id
stringlengths
4
10
text
stringlengths
4
2.14M
source
stringclasses
2 values
created
timestamp[s]date
2001-05-16 21:05:09
2025-01-01 03:38:30
added
stringdate
2025-04-01 04:05:38
2025-04-01 07:14:06
metadata
dict
2203557149
Project teams on Github It seems Github changed something, and I am no longer able to administer teams being a project administrator. I am assuming this is due to some change on Github's part, since I checked with the pypa/build teams and I am getting the same result as in pyproject-metadata teams. Previously I was able to manage team members. @di or @pradyunsg, could you add the following users to the pypa/python-pyproject-metadata-committers team? henryiii rgommers frostming @FFY00 I've invited those folks to that team and also made you a maintainer of all teams for this project as well, so you should be able to make changes in the future.
gharchive/issue
2024-03-23T00:19:16
2025-04-01T04:35:36.378898
{ "authors": [ "FFY00", "di" ], "repo": "pypa/pyproject-metadata", "url": "https://github.com/pypa/pyproject-metadata/issues/92", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1926279727
Add a GitHub Action to run nox and pytest on current Python Run nox on every pull request on Python v3.8, v3.10, and v3.12. Legacy Python died 1,372 days ago on 1/1/2020 so GitHub Actions no longer supports running it. Test results: https://github.com/cclauss/sample-namespace-packages/actions I've also gone ahead and enabled branch protections, and marked these as required. Argh, these checks seem to be failing on the master. Argh, these checks seem to be failing on the master. I'm looking into it. @pradyunsg The errors are related to setup.py installs. After some hacking I found out we're affected by pypa/setuptools#2059 in a special way: The path to the generated .egg gets written to an easy-install.pth in the site-packages directory, but that file is somehow not read, therefore not added to sys.path, which results in Python being unable to find the module. I think the least expensive way to resolve this would be to remove setup.py from the install_commands in the test workflow, for pip works fine: https://github.com/pypa/sample-namespace-packages/blob/b0a98d07a44c94e2e997ee33e71b9bf8fe1b423d/noxfile.py#L31-L32 Are these expected to fail?? Calling setup.py directly should've been wiped out long ago. That invocation method is deprecated. Also, this PR adds updating pip and setuptools, which is absolutely unnecessary with Nox. Thanks for the hints, I'll prepare the follow-up! Calling setup.py directly should've been wiped out long ago. That invocation method is deprecated. Yes, but I thought this was kept for increased coverage. Please see #24 for the improvements. Thanks for the hints, I'll prepare the follow-up! Calling setup.py directly should've been wiped out long ago. That invocation method is deprecated. Yes, but I thought this was kept for increased coverage. That's probably legacy. Also, the only reason setuptools dep was there is to call that setup.py in the outer env too.
gharchive/pull-request
2023-10-04T13:55:48
2025-04-01T04:35:36.386457
{ "authors": [ "cclauss", "chrysle", "pradyunsg", "webknjaz" ], "repo": "pypa/sample-namespace-packages", "url": "https://github.com/pypa/sample-namespace-packages/pull/23", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1173662101
[BUG] version must be imported first to be used as attr setuptools version 60.10.0 Python version 3.9.10 OS Ubuntu 22.04 Additional environment information No response Description (The reproduce this bug, here is an MWE.) In a Python project organized like . ├── pyproject.toml ├── README.md ├── setup.cfg └── src └── foobar ├── __about__.py ├── __init__.py └── _main.py with __about__.py reading __version__ = "1.0.0" I would like to use this __version__ in setup.cfg. I specified version = attr: foobar.__about__.__version__ However, this only works if __init__.py imports __about__ before anything else. In the MWE, a third-party library, rich, is imported via main first, and one gets ModuleNotFoundError: No module named 'rich' when trying to build the project python3 -m build --sdist It's weird that the import order in __init__.py should make a difference here. Expected behavior Build correctly independently of the import order. How to Reproduce See above. Output ModuleNotFoundError: No module named 'rich' Hi @nschloe, thank you very much for reporting this problem. I have the impression that this issue is solved by the improvements in the experimental/support-pyproject branch (planed to be merged soon). I tested your example locally and it seems to be working for that branch. If you have time and would like to test your example and original project against this branch, that would be very helpful! You can do that by editing your pyproject.toml: # replace requires = ["setuptools>=42", "wheel"] # with requires = [ "setuptools @ git+https://github.com/pypa/setuptools@experimental/support-pyproject", ] (There are more information about other features implemented in this branch in https://discuss.python.org/t/help-testing-experimental-features-in-setuptools/13821) Related question: did this approach used to work in previous versions of setuptools? pyproject files are a different game entirely of course. I'll hold my breath. Hi @nschloe you don't have to use pyproject.toml metadata to benefit from the experimental branch 😀 On the process of supporting pyproject.toml metadata, I refactored and improved the code supporting setup.cfg (so the same logic could be shared between both formats). If you want to test if the changes work for you, you just need to change the build-system.requires as I commented above. No need to migrate away from setup.cfg for now. Alright, I tested it with the experimental branch and setup.cfg and can confirm that it fixes the bug. I won't use it in production yet (it's experimental after all) but here's a +1 for merging. :)
gharchive/issue
2022-03-18T14:35:15
2025-04-01T04:35:36.395504
{ "authors": [ "abravalheri", "nschloe" ], "repo": "pypa/setuptools", "url": "https://github.com/pypa/setuptools/issues/3176", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1733815186
Add full content on blog RSS I make an experiment on changing the configurations of RSS feed generation to include the whole content of the posts (#13812). Apparently the pretty_print: true was the responsible for weird RSS. It looks like the results are fine with my newsreader (NetNewsWire): <?xml version="1.0" encoding="UTF-8"?> <rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/"> <channel> <title>The Python Package Index</title> <description>The official blog of the Python Package Index</description> <link>http://0.0.0.0:7000/</link> <atom:link href="http://0.0.0.0:7000/feed_rss_created.xml" rel="self" type="application/rss+xml" /> <docs>https://github.com/pypi/warehouse</docs> <language>en-None</language> <pubDate>Wed, 31 May 2023 09:38:11 -0000</pubDate> <lastBuildDate>Wed, 31 May 2023 09:38:11 -0000</lastBuildDate> <ttl>1440</ttl> <generator>MkDocs RSS plugin - v1.7.0</generator> <image> <url>https://blog.pypi.org/assets/logo.png</url> <title>The Python Package Index</title> <link>http://0.0.0.0:7000/</link> </image> <item> <title>Reducing Stored IP Data in PyPI</title> <author>Mike Fiedler</author> <category>security</category> <category>transparency</category> <description>&lt;div class=&#34;blogging-tags-grid&#34;&gt; &lt;a href=&#34;http://0.0.0.0:7000/tags#security&#34; class=&#34;blogging-tag&#34;&gt;&lt;code&gt;#security&lt;/code&gt;&lt;/a&gt; &lt;a href=&#34;http://0.0.0.0:7000/tags#transparency&#34; ... [cut] ... for reading!&lt;/p&gt;&lt;hr&gt;&lt;p&gt;&lt;em&gt;Mike Fiedler is a PyPI administratorand maintainer of the Python Package Index since 2022.&lt;/em&gt;&lt;/p&gt;</description> <link>http://0.0.0.0:7000/posts/2023-05-26-reducing-stored-ip-data/</link> <pubDate>Fri, 26 May 2023 15:00:00 +0000</pubDate> <source url="http://0.0.0.0:7000/feed_rss_created.xml">The Python Package Index</source> <guid isPermaLink="true">http://0.0.0.0:7000/posts/2023-05-26-reducing-stored-ip-data/</guid> </item> <item> <title>Securing PyPI accounts via Two-Factor Authentication</title> <author>Donald Stufft</author> <category>2fa</category> <category>security</category> <description>&lt;div class=&#34;blogging-tags-grid&#34;&gt; &lt;a href=&#34;http://0.0.0.0:7000/tags#security&#34; class=&#34;blogging-tag&#34;&gt;&lt;code&gt;#security&lt;/code&gt;&lt;/a&gt; &lt;a href=&#34;http://0.0.0.0:7000/tags#2fa&#34; ... [cut] ... denying them access to their account.[^3]: For end users it forces them to purchase some kind of hardware token &lt;em&gt;OR&lt;/em&gt; to use some sort of TOTP application. In both cases it forces them to keep track of something else besides their password and changes the login flow from what they are used to. For PyPI it increases the chance that someone may get locked out of their account, requiring intervention by administrators.[^4]: Not for nothing, but PyPI is also an Open Source project, run largely by volunteers, and cleaning up after a compromise on PyPI is something that affects those volunteers significantly.&lt;/p&gt;</description> <link>http://0.0.0.0:7000/posts/2023-05-25-securing-pypi-with-2fa/</link> <pubDate>Thu, 25 May 2023 00:00:00 +0000</pubDate> <source url="http://0.0.0.0:7000/feed_rss_created.xml">The Python Package Index</source> <guid isPermaLink="true">http://0.0.0.0:7000/posts/2023-05-25-securing-pypi-with-2fa/</guid> </item> <item> <title>PyPI was subpoenaed</title> <author>Ee Durbin</author> <category>compliance</category> <category>transparency</category> <description>&lt;div class=&#34;blogging-tags-grid&#34;&gt; &lt;a href=&#34;http://0.0.0.0:7000/tags#transparency&#34; class=&#34;blogging-tag&#34;&gt;&lt;code&gt;#transparency&lt;/code&gt;&lt;/a&gt; &lt;a href=&#34;http://0.0.0.0:7000/tags#compliance&#34; class=&#34;blogging-tag&#34;&gt;&lt;code&gt;#compliance&lt;/code&gt;&lt;/a&gt; ... [cut] ... QUESTION}&#39;;&lt;/code&gt;&lt;/p&gt;&lt;hr&gt;&lt;p&gt;&lt;em&gt;Ee Durbin is the Director of Infrastructure atthe Python Software Foundation.They have been contributing to keeping PyPI online, available, andsecure since 2013.&lt;/em&gt;&lt;/p&gt;</description> <link>http://0.0.0.0:7000/posts/2023-05-24-pypi-was-subpoenaed/</link> <pubDate>Wed, 24 May 2023 13:12:00 +0000</pubDate> <source url="http://0.0.0.0:7000/feed_rss_created.xml">The Python Package Index</source> <guid isPermaLink="true">http://0.0.0.0:7000/posts/2023-05-24-pypi-was-subpoenaed/</guid> </item> <item> <title>Removing PGP from PyPI</title> <author>Donald Stufft</author> <category>security</category> <description>&lt;div class=&#34;blogging-tags-grid&#34;&gt; &lt;a href=&#34;http://0.0.0.0:7000/tags#security&#34; class=&#34;blogging-tag&#34;&gt;&lt;code&gt;#security&lt;/code&gt;&lt;/a&gt; &lt;/div&gt;&lt;style&gt; .md-typeset .blogging-tags-grid { display: flex; ... [cut] ... was present but had since expired.[^3]: We use meaningfully verified to mean that the signature was valid and the key that made it was not expired and had binding identify information that could tell us if this key was the correct key.&lt;/p&gt;</description> <link>http://0.0.0.0:7000/posts/2023-05-23-removing-pgp/</link> <pubDate>Tue, 23 May 2023 00:00:00 +0000</pubDate> <source url="http://0.0.0.0:7000/feed_rss_created.xml">The Python Package Index</source> <guid isPermaLink="true">http://0.0.0.0:7000/posts/2023-05-23-removing-pgp/</guid> </item> </channel> </rss> Looks like the built RSS has some issues, possibly due to the way we author posts. See https://validator.w3.org/feed/check.cgi?url=https%3A%2F%2Fblogpypiorg--13815.org.readthedocs.build%2Ffeed_rss_created.xml Thanks @osantana !
gharchive/pull-request
2023-05-31T09:56:27
2025-04-01T04:35:36.414469
{ "authors": [ "miketheman", "osantana" ], "repo": "pypi/warehouse", "url": "https://github.com/pypi/warehouse/pull/13815", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
462161923
wktext PROJ parameter being removed during conversion I'm updating some code to work better with pyproj 2.0+ and have run in to a small issue. My goal is to make predefined CRS objects available to Satpy/Pyresample users that work with Proj/CRS from pyproj and CRS from rasterio. The problem projection is: proj_dict = {'proj': 'ob_tran', 'o_proj': 'eqc', 'o_lat_p': 30.0, 'o_lon_p': 10.0, 'lon_0': -10.0, 'a': 6371000.0, 'b': 6371000.0, 'wktext': True} In the past I never used the wktext but found that it was necessary to get rasterio to understand it after doing some googling. Now, I can't find the page that suggested it. I have no idea what it is actually doing. However, when using pyproj's CRS object as an intermediate object the wktext parameter seems to get lost: from pyproj import CRS from rasterio.crs import CRS as rioCRS CRS.from_dict(proj_dict).to_dict() # {'proj': 'ob_tran', 'o_proj': 'eqc', 'o_lat_p': 30.0, 'o_lon_p': 10.0, 'lon_0': -10.0, 'a': 6371000.0, 'b': 6371000.0, 'type': 'crs'} rioCRS.from_dict(proj_dict) # CRS.from_wkt('PROJCS["unnamed",GEOGCS["unnamed ellipse",DATUM["unknown",SPHEROID["unnamed",6371000,0]],PRIMEM["Greenwich",0],UNIT["degree",0.0174532925199433]],PROJECTION["custom_proj4"],EXTENSION["PROJ4","+proj=ob_tran +o_proj=eqc +o_lat_p=30.0 +o_lon_p=10.0 +lon_0=-10.0 +a=6371000.0 +b=6371000.0 +wktext=True"]]') rioCRS.from_dict(CRS.from_dict(proj_dict).to_dict()) # CRSError: The PROJ4 dict could not be understood. OGR Error code 5 Could someone explain the purpose of the wktext parameter and whether or not you would consider this a bug in pyproj or in rasterio or neither? Is there something else I can do to make this projection understood to both pyproj and rasterio CRS? I am thinking 'type': 'crs' is probably causing you issues here. It is new and probably not supported by rasterio. Nope, rasterio seems ok with it: new_dict = CRS.from_dict(proj_dict).to_dict() # {'proj': 'ob_tran', 'o_proj': 'eqc', 'o_lat_p': 30.0, 'o_lon_p': 10.0, 'lon_0': -10.0, 'a': 6371000.0, 'b': 6371000.0, 'type': 'crs'} new_dict['wktext'] = True rioCRS.from_dict(new_dict) # CRS.from_wkt('PROJCS["unnamed",GEOGCS["unnamed ellipse",DATUM["unknown",SPHEROID["unnamed",6371000,0]],PRIMEM["Greenwich",0],UNIT["degree",0.0174532925199433]],PROJECTION["custom_proj4"],EXTENSION["PROJ4","+proj=ob_tran +o_proj=eqc +o_lat_p=30.0 +o_lon_p=10.0 +lon_0=-10.0 +a=6371000.0 +b=6371000.0 +wktext=True"]]') I am not familiar with the reasoning in rasterio for needing wktext. I would post about this issue in rasterio or ask in https://rasterio.groups.io/g/main to get better clarity on why it is needed. As a workaround: >>> crs = CRS.from_dict(proj_dict) >>> rcrs = RCRS.from_wkt(crs.to_wkt("WKT1_GDAL")) >>> rcrs CRS.from_wkt('PROJCS["unknown",GEOGCS["unknown",DATUM["unknown",SPHEROID["unknown",6371000,0]],PRIMEM["Greenwich",0,AUTHORITY["EPSG","8901"]],UNIT["degree",0.0174532925199433,AUTHORITY["EPSG","9122"]]],PROJECTION["custom_proj4"],UNIT["metre",1,AUTHORITY["EPSG","9001"]],AXIS["Easting",EAST],AXIS["Northing",NORTH],EXTENSION["PROJ4","+proj=ob_tran +o_proj=eqc +o_lat_p=30.0 +o_lon_p=10.0 +lon_0=-10.0 +a=6371000.0 +b=6371000.0"]]') Looks like this works as well: >>> rcrs = RCRS.from_wkt(crs.to_wkt()) >>> rcrs CRS.from_wkt('PROJCRS["unknown",BASEGEOGCRS["unknown",DATUM["unknown",ELLIPSOID["unknown",6371000,0,LENGTHUNIT["metre",1,ID["EPSG",9001]]]],PRIMEM["Greenwich",0,ANGLEUNIT["degree",0.0174532925199433],ID["EPSG",8901]]],CONVERSION["unknown",METHOD["PROJ ob_tran o_proj=eqc"],PARAMETER["o_lat_p",30,ANGLEUNIT["degree",0.0174532925199433,ID["EPSG",9122]]],PARAMETER["o_lon_p",10,ANGLEUNIT["degree",0.0174532925199433,ID["EPSG",9122]]],PARAMETER["lon_0",-10,ANGLEUNIT["degree",0.0174532925199433,ID["EPSG",9122]]]],CS["Cartesian",2],AXIS["(E)",east,ORDER[1],LENGTHUNIT["metre",1,ID["EPSG",9001]]],AXIS["(N)",north,ORDER[2],LENGTHUNIT["metre",1,ID["EPSG",9001]]]]') Yeah I can try that, but still hard to use. I will point the rasterio people to this issue to see if they have any ideas. @snowman2 Any idea why pyproj/PROJ drops the +wktext? Is there a rule somewhere that says "this shouldn't be used any more"? I did a quick search in the PROJ code and didn't find wktext anywhere. I also didn't see it in the code for rasterio. Maybe it is specific to GDAL? For pyproj, it just uses the input string to create a projection object. Then on export, it converts that object back out to a string. Since wktext is not in the PROJ codebase as far as I can tell, I assume it just gets dropped. Yep, looks like it is an OGR specific thing: https://github.com/OSGeo/gdal/blob/9f15a2e5f5b15e1eb5a927d8496c23d0d4c3f861/gdal/ogr/ogrspatialreference.cpp#L9508-L9516 Interesting. That makes sense. It looks like the wktext is preserved in the srs or user input string of the pyproj object: >>> crs = CRS({'proj': 'ob_tran', 'o_proj': 'eqc', 'o_lat_p': 30.0, 'o_lon_p': 10.0, 'lon_0': -10.0, 'a': 6371000.0, 'b': 6371000.0, 'wktext': True}) >>> crs.to_string() '+proj=ob_tran +o_proj=eqc +o_lat_p=30.0 +o_lon_p=10.0 +lon_0=-10.0 +a=6371000.0 +b=6371000.0 +wktext +type=crs' I am wondering if adding a kwarg to to_proj/to_dict to help preserve this may help or even makes sense to do. For example, something like: >>> crs.to_dict(from_input=True) {'proj': 'ob_tran', 'o_proj': 'eqc', 'o_lat_p': 30.0, 'o_lon_p': 10.0, 'lon_0': -10.0, 'a': 6371000.0, 'b': 6371000.0, 'type': 'crs', '+wktext': True} My main concern with this idea is that it has the potential to propagate potential invalid arguments that users pass in. Due to this. users might assume that they are valid or supported PROJ arguments. But, then again, it is returned on to_string. :thinking: I am personally not sure we should add such a keyword. As you said, it propagates invalid arguments that are not understood by PROJ anyway. I am not sure we should therefore keep them in a to_proj. It seems that Sean suggested to use the WKT anyway to interact between pyproj and rasterio, instead of the proj dict (so similar as your suggestion above). @djhoese that doesn't solve the issue? (you mentioned above that this is hard to use, can you clarify that a bit more?) I think I agree that pyproj probably shouldn't change anything the more I learn about this parameter. I think @snowman2's suggestion will work fine, but this if for the pyresample library which still depends pretty heavily on the PROJ dict. This issue was discovered in my first attempts at switching to the CRS object internally, but still providing access to a PROJ dict/str property in pyresample. I can see two easy-ish solutions for me: Update the rasterio-based code in my libraries to use the CRS WKT if it exists. WKT is able to fully describe all CRSes right? Add +wktext for ob_tran and nsper projections whenever a PROJ dict is requested. I just checked my post on the rasterio community forum/group and @sgillies suggests the same as @snowman2. That is, use WKT for conversion between pyproj and rasterio and not PROJ. Is there any reason I might use some other WKT version/format instead of pyproj's default of WKT2_2018? WKT is able to fully describe all CRSes right? As far as I know it should. Is there any reason I might use some other WKT version/format instead of pyproj's default of WKT2_2018? You would probably need to check when the support was added in GDAL for WKT2_2018. The recent versions support it, but depending on what version of GDAL/rasterio you would like to support, it may fail. So, the WKT1_GDAL is also an option. I am personally not sure we should add such a keyword. I think if there is a lot of hesitation on something, probably better not. I think removing that sounds like a good idea. I believe this is resolved. If you have further questions or feel it needs to be reopened, feet free to reopen or just post further questions. I need to bring this back up because I ran in to another issue related to converting a pyproj CRS object to a rasterio CRS object. I have PROJ 6.1.1, pyproj 2.3.1, libgdal 2.4.2, and rasterio 1.0.28 all from conda-forge installed. I'm trying to take any pyproj CRS object and make a geotiff with it using rasterio. From the above discussion I did crs = rasterio.crs.CRS.from_wkt(pyproj_crs.to_wkt()) but found that this fails for EPSG:3035: CPLE_AppDefinedError: Only OGC WKT Projections supported for writing to GeoTIFF. PROJCRS["ETRS89 / LAEA Europe",BASEGEOGCRS["ETRS89",DATUM["European Terrestrial Reference System 1989",ELLIPSOID["GRS 1980",6378137,298.257222101,LENGTHUNIT["metre",1]]],PRIMEM["Greenwich",0,ANGLEUNIT["degree",0.0174532925199433]],ID["EPSG",4258]],CONVERSION["Europe Equal Area 2001",METHOD["Lambert Azimuthal Equal Area",ID["EPSG",9820]],PARAMETER["Latitude of natural origin",52,ANGLEUNIT["degree",0.0174532925199433],ID["EPSG",8801]],PARAMETER["Longitude of natural origin",10,ANGLEUNIT["degree",0.0174532925199433],ID["EPSG",8802]],PARAMETER["False easting",4321000,LENGTHUNIT["metre",1],ID["EPSG",8806]],PARAMETER["False northing",3210000,LENGTHUNIT["metre",1],ID["EPSG",8807]]],CS["Cartesian",2],AXIS["northing (Y)",north,ORDER[1],LENGTHUNIT["metre",1]],AXIS["easting (X)",east,ORDER[2],LENGTHUNIT["metre",1]],USAGE[SCOPE["unknown"],AREA["Europe - ETRS89"],BBOX[32.88,-16.1,84.17,40.18]],ID["EPSG",3035]] not supported. If I directly create the rasterio CRS object with the EPSG code then it works fine. If I use version='WKT1_ESRI' or version='WKT1_GDAL' (which I know you mentioned above @snowman2). I couldn't find an exact answer on when WKT2 was added to GDAL but it is confidently in GDAL 3.0+. Any suggestions about what version I should use to export the CRS information to rasterio? Does rasterio just not understand WKT2 and that's why my usage fails? Any help is appreciated. Hmm, sounds like you have done some good investigation so far. I guess you could do a trial and error approach. You could use to_epsg and if that fails, the to_wkt starting with WKT2 and then fall back to the other versions. Yeah the issue is I haven't found a good way to test if the geotiff creation will fail without actually trying to create the geotiff. I think even the rasterio CRS object's wkt property which I think is documented as producing an OGC compatible version of the WKT still returns a string and then geotiff creation will fail. Hmmm, do you have a reproducible example? If so, it might be worth it to raise the issue with rasterio. I do have a reproducible example, but now things are getting even more confusing. Looks like this is only an issue with libgdal <3.0 install (assuming because WKT2 is supported in GDAL). Currently debugging more, will file an issue if I track it down to the actual problem.
gharchive/issue
2019-06-28T18:42:00
2025-04-01T04:35:36.433894
{ "authors": [ "djhoese", "jorisvandenbossche", "snowman2" ], "repo": "pyproj4/pyproj", "url": "https://github.com/pyproj4/pyproj/issues/357", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
130035552
Aggregate metrics lag Aggregate metrics are currently lagging if the sending paty is paused or stops for some period of time. Aggregation engine keeps on moving forward the previous snapshot of the agent instead of using the new snapshot despite, even if time difference if significant. UPD: it turned out that the lag was caused by the processing of the query results rather than their storage. PR that fixes the issue is available.
gharchive/issue
2016-01-30T20:57:45
2025-04-01T04:35:36.445605
{ "authors": [ "ifesdjeen" ], "repo": "pyr/cyanite", "url": "https://github.com/pyr/cyanite/issues/170", "license": "isc", "license_type": "permissive", "license_source": "bigquery" }
131474433
Ring buffer integration Disruptor is known to beat queue performance extremely well. Certainly this prototype doesn't involve things like back pressure, event optimisation (GC pressure because of event instances created every time) also memory access across the barrier and many other possible optimisations, but it might serve as a good basis for further discussion. My current benchmarks show that it handles: java -jar stresser.jar localhost 2003 10 20 1 true without much trouble (10 hosts, 20 timers (x15 metrics per second) = 3000 metrics per second) . Im quite sure we can pull out much more than that. That's a non-optimised no nothing version... @ifesdjeen I moved the protocol on the queues themselves and stored ingestq and writeq directly. I also took the opportunity to remove some functions which are now unused. It seems to make the namsepace overall more legible. Sure, that's a great idea. Cleaning up functions is a good thing as well imo. Looks good to my opinion. @ifesdjeen I'm inclined to go ahead and merge this quickly. To address the points you raised in the initial description, and to understand what follow-up work we will have to do: What are common patterns to avoid the GC pressure you mention, i.e: how would you avoid event instance creation? What happens when capacity is reached on the queue? With the atom gone, what memory access should we watch out for? @ifesdjeen @mpenet I added a commit which uses volatile! thanks for the pointer. What are common patterns to avoid the GC pressure you mention, i.e: how would you avoid event instance creation? In reactor, you can "lend" the objects, modify them and then publish (this is hidden behind the disruptor dsl for now). So in essence, I'd lend a metric object from the queue, set all the required fields there and then publish. This way we avoid recreating it. Of course that'd mean larger initial queue size but in a long term it pays off :) What happens when capacity is reached on the queue? Current behaviour: backpressure (blocking). Although it's configurable and we can either drop events or do something else. With the atom gone, what memory access should we watch out for? with the atom gone it should be good at least here :) i thought we could use something like primitive mutable java wrappers or alike in order to adress make the snapshot more lightweight or even byte buffers but that might be a bit too early to even consider that. @pyr you are faster than the wind today @ifesdjeen ok, thanks for all the information and for putting in the work. this looks awesome. w0000000t :)
gharchive/pull-request
2016-02-04T21:09:40
2025-04-01T04:35:36.452108
{ "authors": [ "ifesdjeen", "pyr" ], "repo": "pyr/cyanite", "url": "https://github.com/pyr/cyanite/pull/174", "license": "isc", "license_type": "permissive", "license_source": "bigquery" }
741955745
[scripts] Add script to create super user Feature Add script to create super user Motivation After a possible deletion of the volume of the db, it would be convenient to have a script to create a super user. Pitch We might add a new folder called scripts with a first one, adding a super user with admin scope. I don't think there will but make sure to check #17 to avoid PR duplication with @florianriche ;) Having one script to create a super user will be useful even in production, and an end-to-end script is more for example and testing purposes I don't know if a script is the right way to init a superuser. In the cookie-cutter of fastapi, there is a piece of code that is executed at launch time to create this superuser. If the script exists, isn't there a risk (albeit small) that the superuser is created multiple times? def init_db(db: Session) -> None: # Tables should be created with Alembic migrations # But if you don't want to use migrations, create # the tables un-commenting the next line # Base.metadata.create_all(bind=engine) user = crud.user.get_by_email(db, email=settings.FIRST_SUPERUSER) if not user: user_in = schemas.UserCreate( email=settings.FIRST_SUPERUSER, password=settings.FIRST_SUPERUSER_PASSWORD, is_superuser=True, ) user = crud.user.create(db, obj_in=user_in) # noqa: F841``` Up to discuss Thanks for the feed back. Indeed, the possibility to create at launch time might be a better way ! We would run init_db, similar function to the one in the cookie-cutter, just after database connection in main.py (and having the init_db func somewhere)? or, should we have a separate script for inits, a bit like in cookie cutter (https://github.com/tiangolo/full-stack-fastapi-postgresql/blob/master/{{cookiecutter.project_slug}}/backend/app/app/initial_data.py) and run it in adding a command in the docker-compose file ? I agree with you, having a script could led to have multiples super user. Indeed, if it's possible to do this at initialization, that would be perfect! Having a separate python module like in the cookie cutter would be good option I think :ok_hand:
gharchive/issue
2020-11-12T21:56:21
2025-04-01T04:35:36.470259
{ "authors": [ "fe51", "florianriche", "frgfm" ], "repo": "pyronear/pyro-api", "url": "https://github.com/pyronear/pyro-api/issues/26", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1243086568
feat: Added codecarbon integration This PR introduces the following modifications: adds decorator integration of codecarbon updates requirements and gitignore Any feedback is welcome! cc @benoit-cty Hi @frgfm thanks for this ! I do it work ? do we have an associated login ? or anyone cloning this repo will impact our score ? I think it's better not to track public users and to offer access to anyone who is close to the project and making them aware of gpu abuse. Hello, Your current code will work with no config file : the user emissions will be tracked but not send to the CodeCarbon API. Your user will have a local file emissions.csv. It's the default behavior of CodeCarbon. So I think you are fine: When you are in an environment with proper config file, your emissions will go into the CodeCarbon API. In another PR, we can throw a warning when there is no config file (so that core team doesn't forget to set it when running a training)
gharchive/pull-request
2022-05-20T12:33:57
2025-04-01T04:35:36.473523
{ "authors": [ "MateoLostanlen", "benoit-cty", "frgfm" ], "repo": "pyronear/pyro-vision", "url": "https://github.com/pyronear/pyro-vision/pull/140", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
341320705
Printing to stdout during compilation should be avoided rust-embed prints to stdout during compilation: https://github.com/pyros2097/rust-embed/blob/2025bfb7b8becb4f281efa5cd3febc611ba98fcb/src/lib.rs#L56 macros should not do that. It breaks editors that parse the json output: https://github.com/rust-lang/rust-enhanced/issues/310#issuecomment-405052274 All occurrences of println and eprintln in that file should be removed. Well.. I thought it would be good to see which files are being loaded into the binary at compile time. That would make it easy to find out the progress of the compilation and also if some large file is taking too long. Does printing to stderr is a better approach or removing it altogether better. Anyways I'm ok with removing the println no one else has issues with it. But will be :sob: as I have to remove the cool screenshots of my terminal from the readme.md The include_bytes!() call should be very fast though, right? It's not really "compiling" (processing) the included files.. So I think it's ok to have no output about this. People who are using rust-embed will know that if it takes a bit longer, it's because of the embedded files :) I think a better solution is to just mention it in the Readme: "Be aware that compilation might take a little bit longer when embedding a large number of files." Ok. You convinced me. And it seems no one is going to miss it. Do you want to raise a PR for it or should i do it. I'm currently traveling on vacation with only my phone, so I think it'd be faster if you do it :)
gharchive/issue
2018-07-15T13:22:29
2025-04-01T04:35:36.477943
{ "authors": [ "Boscop", "pyros2097" ], "repo": "pyros2097/rust-embed", "url": "https://github.com/pyros2097/rust-embed/issues/27", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
2401560858
removing spvcm from meta release -- archived - #1330 This PR removing spvcm from meta release, as it is now archived - https://github.com/pysal/spvcm/issues/18 resolves #1330 I'm not sure whether we deprecated spvcm before, so removing it from meta may be pretty drastic in case we haven't. If we haven't, we should deprecate spvcm during the 24.07 meta cycle with documentation, then remove in 25.01. dunno either, but also not sure how would we issue a deprecation warning in meta? OK, so I shouldn't have arched it like we discussed? Should I "unarchive"? OK, so I shouldn't have arched it like we discussed? Should I "unarchive"? No, archiving is fine, I don't think we need to revert that. I'm wondering about the best way to surface the archiving in the meta release. What about updating the meta README.md like: Works for me. So we would keep this pr open until 25.01 release, but I can add the warning to the docs in the 24.07 release. So we would keep this pr open until the 25.01 release, but I can add the warning to the docs in the 24.07 release. Probably the best idea. --> Should also post a notice like I suggest in https://github.com/pysal/pysal.github.io/issues/229?
gharchive/pull-request
2024-07-10T19:56:31
2025-04-01T04:35:36.486293
{ "authors": [ "jGaboardi", "knaaptime", "sjsrey" ], "repo": "pysal/pysal", "url": "https://github.com/pysal/pysal/pull/1337", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
459997296
Message.edit_attach bugfix Updated Message.edit_attach to work with inline keyboards I had to force-push because this repo requires signing all commits
gharchive/pull-request
2019-06-24T16:57:51
2025-04-01T04:35:36.583327
{ "authors": [ "MarcoBuster" ], "repo": "python-botogram/botogram", "url": "https://github.com/python-botogram/botogram/pull/127", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1430868746
Radicale uses nonstandard URLs It appears that Radicale is using non standard urls to host caldav files. On normal caldav servers, (nextcloud, etc), the caldav file can be found at the following location example.com/user/03898134-59a0-11ed-a6d7-cb4c1cf0d9f0 but under radicale it can be found at example.com/user/tasks-03898134-59a0-11ed-a6d7-cb4c1cf0d9f0 (notice the tasks- bit) or example.com/user/calendars-03898134-59a0-11ed-a6d7-cb4c1cf0d9f0 (notice the calendars- bit) When creating a calendar on a radicale server using python-caldav, the url of the generated calendar does not reflect radicale's actual url format and that results in 404 errors when trying to modify the calendar. I'm a bit confused on this one. Radicale is one of the two calendar servers that I'm testing frequently against, so at least all the test code passes. If you have any troubles with radicale, then I encourage you to write up some specific code that fails. The caldav standard puts no constraint on the file name of the event/task/journal, but it should be a direct child of the calendar if I've understood it correct. So if the calendar URL is example.com/user/, then both example.com/user/calendars-03898134-59a0-11ed-a6d7-cb4c1cf0d9f0.ics and example.com/user/03898134-59a0-11ed-a6d7-cb4c1cf0d9f0.ics should be valid URLs. There need not be any correlation between the URL/filename and the UID of the event, but the caldav library will use the uid as the filename for new events. For a new event with UID 03898134-59a0-11ed-a6d7-cb4c1cf0d9f0, the caldav client library will use example.com/user/03898134-59a0-11ed-a6d7-cb4c1cf0d9f0.ics as the URL. It's a while since last time I read the RFC, but I believe the calendar server may rename the file and return a 301 or 302 redirect. The caldav client library will honor such a redirect. The caldav standard also does not put many restrictions on the UID. Both 03898134-59a0-11ed-a6d7-cb4c1cf0d9f0 and tasks-03898134-59a0-11ed-a6d7-cb4c1cf0d9f0 are valid UIDs. The caldav standard encourages the usage of timestamp, serial number and domain in the UID - like 20010712T182145Z-123401@example.com. At some point I decided to use uuids rather than the format above when generating UIDs, perhaps I didn't read the above well enough. Thanks, I'll have a look into it when I get time. The children method is one out of several ways to get things out of the calendar. Dependent on how well the server supports the standards, some methods may work on some servers and others not. Instead of calendar.children() it's also possible to try to use calendar.objects() or one or more of those: calendar.events() ,calendar.todos(), calendar.journals(). Even calendar.search() may be used. It's a bit messy, I intend to make a more tidy interface in 1.0, as well as ways of discovering or configuring compatibility problems and working around those. When calendar.children() works for all calendars except those that are created through the library, this sounds like a problem at the library side, not the server side. I will try to copy your code above into the testing framework and take it from there. I'm quite busy this week and probably into the next one as well, not sure when I will find a time slot for this. Right, all those methods fail with the above error. In fact, any method that attempts to draw data from the server results in that 404 error because the URL that pycaldav sets when creating the calendar is not the URL that one would receive from running principal.calendars() pycaldav when creating the calendar sets the calendar.url to something like this example.com/user/03898134-59a0-11ed-a6d7-cb4c1cf0d9f0 but the url of the calendar that actually gets created on the radicale server looks like this example.com/user/tasks-03898134-59a0-11ed-a6d7-cb4c1cf0d9f0 with either tasks- or calendar- preceeding the ID, depending on the type of object that was created. pycaldav is attempting to access the server with the wrong URL resulting in the 404 not found errors. Cool. I'll ensure we have tests and a fix in the upcoming 0.10.1-release. I'm not able to reproduce. What version of Radicale do you use? 3.1.8 (from PyPI) Still not able to reproduce. Can you try to download the file I just made - https://github.com/python-caldav/caldav/blob/ece0d29adb762ee381368214ccc5115ada1ece45/tests/tmp_226.py - and see if it does the same in your environment as mine? Your script works but the problem still occurs with my radicale install. The only difference i can determine is that radicale is being run from within python in your script's case while in my script's case its a seperate process I'm also not able to reproduce when starting up radicale as a separate process Perhaps you have some site-local configuration causing this behaviour? Anyway, looking into radicale/app/mkcalendar.py in the radicale project, I don't see any code prepending the calendar path in any way. I can see one potential bug in the current client side code - it will ignore a 301/302 redirect from the server side - it should probably be picked up and interpreted as the correct location for the calendar. Perhaps fixing this would solve your problem ... but I would be sort of working blindfolded if I was to implement anything like that. I would probably either need access to run the test code towards your server, or you would have to do the work yourself. It's in caldav/object.py, class Calendar, def _create, after r = self._query there should be a check for r.status, and if it's 301/302, then self.url needs to be changed. Here's my config file: [storage] type = radicale_storage_decsync filesystem_folder = <path to filesystem folder> decsync_dir = <path to decsync dir> [auth] type = htpasswd htpasswd_filename = <path to password file> # encryption method used in the htpasswd file htpasswd_encryption = md5 [server] hosts = 0.0.0.0:5232, [::]:5232 Do you have the caldav/object.py folder with the changes i could test? [storage] type = radicale_storage_decsync From the config file template, the valid options here are multifilesystem and multfilesystem_nolock. While I find lots of references in the code for multifilesystem, I find no references for radicale_storage_decsync. However, I do find some references to radicale_storage_decsync using a popular search engine ... https://github.com/39aldo39/Radicale-DecSync ... I think that this is not a part of Radicale. I guess that this is the reason why the calendar URL changes after saving it. It appears that the radicale storage is indeed the cause of the issue. Closing ticket.
gharchive/issue
2022-11-01T04:54:45
2025-04-01T04:35:36.598276
{ "authors": [ "BeatLink", "tobixen" ], "repo": "python-caldav/caldav", "url": "https://github.com/python-caldav/caldav/issues/226", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
687246335
Auto-respond to non-mod everyone pings. The issue: Currently, if someone puts down an @everyone ping, a mod has to come and respond to it to tell the person not to ping 95k+ people. Now, this usually works fine, but accidents happen, and things can end up happening like lemon accidentally pinging everyone, because android doesn't have a check for that (we're not upset lemon!). Proposed solution: We add a auto-respond to the bot, sort of like in the attachment auto-respond, that contacts the user to explain why they shouldn't ping 95k people, and maybe even temp-mutes them and/or deletes the message. Here is an example message: Sorry @user, but please don't try to ping 95k people. It will not have good results. If you want to see what can happen, take a look at this pin: https://discordapp.com/channels/267624335836053506/267624335836053506/748352909904969740 It should only respond to non-staff too. This should be implemented as a filter. Automatically deletes the message with the fake everyone ping in it, because we don't want any idiot owners quoting those messages by accident :grin: Mutes the user for 10 minutes, similar to what we do for spam detectors. Posts a message to the channel with a ping of the user and a message similar to what's in OPs description. Maybe we don't need to link the pinned message, lol. I just thought linking the pinned message would help people realize what happens when more 95k+ people get pinged :lemon_pleased:. I think this link is not good idea, but yes, this should have filter and I can create it, when someone else don't do this. The link was more of a joke, but I could also implement this to the bot. I think since you created the issue, @MrAwesomeRocks, you can handle it if you want it. @lemonsaurus How does the current spam filter work? Could you send one of the current filter files? See this cog: https://github.com/python-discord/bot/blob/master/bot/cogs/antispam.py And these rules: https://github.com/python-discord/bot/tree/master/bot/rules @lemonsaurus There's a section in the config file talking about ping everyone, is this implemented anywhere? So I have to: Add my rule to the config file. Write my rule to: Get relevant messages Read the everyone role from the config file Check if the everyone role is mentioned in the config file Add my rule to this dict @lemonsaurus
gharchive/issue
2020-08-27T13:32:30
2025-04-01T04:35:36.606918
{ "authors": [ "MrAwesomeRocks", "ks129", "lemonsaurus" ], "repo": "python-discord/bot", "url": "https://github.com/python-discord/bot/issues/1131", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
721531699
Package cmd closes #796 this command take package name as input and return data from pypi api. package command will return this if package is found package command will return this if package is not found [x] Join the Python Discord Community? [x] If dependencies have been added or updated, run pipenv lock? [x] Lint your code (pipenv run lint)? [x] Set the PR to allow edits from contributors? @Anubhav1603 will you be finishing off this feature? I personally would like to see it get implemented. @Anubhav1603 will you be finishing off this feature? I personally would like to see it get implemented. i am sick these days cant focus on coding can i get some time to work i am sick these days cant focus on coding can i get some time to work @Anubhav1603 yeah, you can get more time. If you decide to drop this, let me know. I hope you get well soon! :heart: @Anubhav1603 yeah, you can get more time. If you decide to drop this, let me know. I hope you get well soon! :heart: @Anubhav1603 Greetings. I hope you've recovered and are feeling better. Will you be willing to pick this PR back up? If not, I can take over for you. sure go a head I'll be doing this on a new PR, then. Thank you for your time, I hope you've gotten better!
gharchive/pull-request
2020-10-14T14:52:48
2025-04-01T04:35:36.612652
{ "authors": [ "Anubhav1603", "Xithrius" ], "repo": "python-discord/bot", "url": "https://github.com/python-discord/bot/pull/1237", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
429866207
Made announcement command and edit command I added this to my bot, so I figured why not add it here. It announces a message and pings the role and allows you to edit the message as well. As mentioned in the last PR, we don't wish to go with this type of implementation. It's not the inability to edit the message that was the issue, but the fact that we don't wish to rely on a bot for every announcement made in the future. Ensuring a human is doing the announcement allows for a more personalised feel for our community announcements, and gives a point of contact if people wish to ask about any specifics about the announcement made. Due to the above reasons, this and the previous PR's are not accepted and won't be part of the possible options that are considered in our next staff meeting that will establish what option we wish to go with. We will decide internally from 5 possible choices that we have gotten already, and implement the decision made ourselves. If this involves a code change to the bot, we will personally do so and have strict reviewing among ourselves during that, so this isn't a thing you should worry about in the future. For now, the owners of the community have already established that anyone who have or will abuse the mentionable roles will end up with a permanent ban. This has already been actioned for any past cases.
gharchive/pull-request
2019-04-05T17:43:19
2025-04-01T04:35:36.615044
{ "authors": [ "ajgrinds", "scragly" ], "repo": "python-discord/bot", "url": "https://github.com/python-discord/bot/pull/345", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1386949451
Allow exporting constraints for pip I've been using poetry export to generate a constraints file for use with pip. Sometime after the poetry 1.2.0a2 and poetry-core 1.1.0a7 releases, this command is now includes extras in its output such as docker[ssh]==6.0.0 which is not supported in pip constraints files. Would you be open to at least adding an option to change this behavior? The easiest to implement would probably be something like a --no-extras flag (which may need a better name to avoid confusion with --extras). Another option would be to add support for a constraints.txt format. We could definitely do this as constraints.txt instead of requirements.txt -- you are welcome to implement a new exporter, but I don't think anyone who works on this repo regularly has expressed any interest in writing it. Thanks for the quick feedback. I'll take a deeper look and see how feasible it'd be for me to write the new exporter myself. My use case is for https://github.com/certbot/certbot. It's a large monorepo containing many Python packages, some of which need to be able to function without the others, and the dependencies vary based on the OS and Python version. To help us pin everything down more reliably while allowing us to install only the subset of packages that's actually needed, we have dummy pyproject.toml files under https://github.com/certbot/certbot/tree/master/tools/pinning and scripts that use poetry to generate and export files which we then use as constraints for pip. Adding a constraints export method to https://github.com/python-poetry/poetry-plugin-export/blob/main/src/poetry_plugin_export/exporter.py doesn't look too bad. Do you know if this text from https://pip.pypa.io/en/latest/user_guide/#constraints-files contains the full list of differences that you'd want? Their syntax and contents is a subset of Requirements Files, with several kinds of syntax not allowed: constraints must have a name, they cannot be editable, and they cannot specify extras. I'm imagining just dropping any extras declared for the package. For the two conditions of "constraints must have a name" and "they cannot be editable", I'd either error out or exclude the package based on y'all's preference. I personally would factor out exporter.py into an abstract base class, then add two concrete implementations that set booleans on the base exporter. The base exporter could then output slightly differently based on the booleans set by each implementation. You have found the canonical source for the format -- I would suggest failing for now myself on any unexportable constraints. Thanks for the guidance. To be clear, are you OK with the constraints exporter dropping extras (e.g. turning docker[ssh]==6.0.0 into docker==6.0.0) instead of erroring out? This behavior is what I personally want for my use case but I also think it makes sense for a constraints format. I think it only makes sense for extras, as by definition constraints is 'versions without install details.' The same packages will be captured as Poetry locks all requested extras. Editable installs are ambiguous and I think we should error for them. You won't need to deal with unnamed constraints as the exporter already names them with <name> @ <url>. Looking at the code instead of remembering what it used to look like, it looks like we've already got support for format-based dynamic dispatch. As such I'd suggest that you simply move the body of _export_requirements_txt into a new method that takes some boolean values, rewrite the method _export_requirements_txt to invoke the new common method, and introduce _export_constraits_txt using the existing dynamic dispatch. This also came up in #108 Not to discourage contributions - quite the opposite, this is evidence that there's real demand for this! - but as I suggested there, you might find it considerably less effort to sed the extras away so that the exported requirements.txt works for you... I don't disagree @dimbleby, but I do think that given certbot does seem to need 'real' constraints, first class support is a good idea so that they don't break in the future because of a brittle reliance on sed/"don't set develop = true". I'm absolutely not against Someone adding this function - as I say, evidence is mounting that it's a thing people want. I do think it's only fair to be sure that OP is aware that they probably have quicker (and dirtier) solutions available before they spend a bunch of time being that Someone. I opened https://github.com/python-poetry/poetry-plugin-export/pull/128 for this.
gharchive/issue
2022-09-27T01:42:17
2025-04-01T04:35:36.694318
{ "authors": [ "bmw", "dimbleby", "neersighted" ], "repo": "python-poetry/poetry-plugin-export", "url": "https://github.com/python-poetry/poetry-plugin-export/issues/125", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2609910868
🛑 Billing Api is down In d9c66e7, Billing Api (https://billing.roblox.com/) was down: HTTP code: 404 Response time: 57 ms Resolved: Billing Api is back up in 8010c94 after 11 minutes.
gharchive/issue
2024-10-23T21:46:33
2025-04-01T04:35:36.873183
{ "authors": [ "pythoniaweb" ], "repo": "pythoniaweb/statusboom", "url": "https://github.com/pythoniaweb/statusboom/issues/1144", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2613031425
🛑 Billing Api is down In aae9213, Billing Api (https://billing.roblox.com/) was down: HTTP code: 404 Response time: 36 ms Resolved: Billing Api is back up in d301bae after 1 hour, 40 minutes.
gharchive/issue
2024-10-25T04:18:19
2025-04-01T04:35:36.875504
{ "authors": [ "pythoniaweb" ], "repo": "pythoniaweb/statusboom", "url": "https://github.com/pythoniaweb/statusboom/issues/1243", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
619391632
Java 14 upgrade Build new version with Java 14. It's impossible with Google's JIB
gharchive/issue
2020-05-16T06:19:36
2025-04-01T04:35:36.877437
{ "authors": [ "pythoninja" ], "repo": "pythoninja/randomicu", "url": "https://github.com/pythoninja/randomicu/issues/99", "license": "Unlicense", "license_type": "permissive", "license_source": "github-api" }
891495424
Indicators I have one question. Shouldn't the indicators be for more days? I mean if you get 1h candles and the RSI is calculated with 14 days, shouldn't it be 24 x 14 = 336 margin rows? Because I understand that the library takes the days as rows of the data frame, as if the rows were days. As far as I know, RSI is calculateb by 14 previous candles. In 1h timeframe, last 14 hourly candles are used. It's worth noting that the value of 14 is a default parameter, but can be adjusted to accommodate different analysis preferences or trading strategies.
gharchive/issue
2021-05-14T00:53:44
2025-04-01T04:35:36.879663
{ "authors": [ "johnnynunez", "kianfa" ], "repo": "pythonlessons/RL-Bitcoin-trading-bot", "url": "https://github.com/pythonlessons/RL-Bitcoin-trading-bot/issues/7", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1456300822
🐛 [Bug] resnet50 model giving error when given dynamic shape input while aten::resize is in fallback Bug Description resnet50 torch model gives an error when given dynamic shape and having resize in fallback mode compiled_cpp_mod = _C.compile_graph(module._c, _parse_compile_spec(spec)) RuntimeError: Trying to create tensor with negative dimension -1: [-1, 3, 224, 224] To Reproduce Steps to reproduce the behavior: Run torch_tensorrt.compile with resnet50 model as input, using fp32 precision and fp32 precision input Put aten::resize in torch_executed_ops Input shapes - shapes = {"input_ids": {"min": (1, 3, 224, 224), "opt": (256, 3, 224, 224), "max": (256, 3, 224,224)}} with max_workspace_size = 8589934592 The above fails for maxine_vfx_MLP_expression as well. For the model, you could mail me. Environment Torch-TensorRT Version: 1.3.0a0+e3b99294 PyTorch Version: 1.13.0.dev20220921+cu116 CPU Architecture: Intel Xeon CPU OS: Ubuntu 18.03 How you installed PyTorch: pip Build command you used: python setup.py develop Are you using local sources or building from archives: local Python version: 3.8.13 CUDA version: 11.6 Build information about Torch-TensorRT can be found by turning on debug messages Torch-TensorRT Version (e.g. 1.0.0): PyTorch Version (e.g. 1.0): CPU Architecture: OS (e.g., Linux): How you installed PyTorch (conda, pip, libtorch, source): Build command you used (if compiling from source): Are you using local sources or building from archives: Python version: CUDA version: GPU models and configuration: Any other relevant information: Additional context I'm unable to reproduce this error. Were you trying with master branch ? Yes this was in the master branch with dynamic shape fallback enabled Can you post the entire script which has the reproducer code ? `import pathlib from typing import Optional from unicodedata import name import numpy as np import torch_tensorrt import torch exported_model_path = "/home/abose/Documents/work/models_exported/models_Bermuda/resnet50-fp16-pytorch/NVIDIA-A100-80GB.nav/torchscript-trace/model.pt" converted_model_path = "/home/abose/Documents/work/models_exported/models_Bermuda/resnet50-fp16-pytorch/NVIDIA-A100-80GB.nav/torch-trt-trace-fp32/model.pt" shapes = {"input_ids": {"min": (1, 3, 224, 224), "opt": (256, 3, 224, 224), "max": (256, 3, 224,224)}} input_dtype = [torch.float32] max_workspace_size = 8589934592 target_device = "cuda" enabled_precisions = {'enabled_precisions': [torch.float32]} Adapted from convert function from reproduce_conversion.py workdir = pathlib.Path.cwd() model_input_shapes = [] for input_shapes, input_dtype in zip(shapes.values(), input_dtype): model_input_shapes.append(torch_tensorrt.Input( min_shape=input_shapes["min"], opt_shape=input_shapes["opt"], max_shape=input_shapes["max"], dtype=input_dtype, ) ) enabled_precisions = {'enabled_precisions': [torch.float32]} exported_model_path = pathlib.Path(exported_model_path) if not exported_model_path.is_absolute(): exported_model_path = workdir model = torch.jit.load(exported_model_path.as_posix(), map_location=target_device) with torch_tensorrt.logging.debug(): tr_model_compiled = torch_tensorrt.compile(module=model, inputs=model_input_shapes, workspace_size=max_workspace_size, truncate_long_and_double=True, torch_executed_ops=["aten::resize"], **enabled_precisions ) converted_model_path = pathlib.Path(converted_model_path) if not converted_model_path.is_absolute(): converted_model_path = workdir / converted_model_path tr_model_compiled.save(converted_model_path.as_posix())` Apurba to run with latest changes and post logs. Posting the logs. The test passes with no dynamic shape. With the latest master changes, the test fails with ERROR: [Torch-TensorRT TorchScript Conversion Context] - 4: [graphShapeAnalyzer.cpp::analyzeShapes::1294] Error Code 4: Miscellaneous (IShuffleLayer %453 : Tensor = aten::reshape(%x.1, %451): reshape dimension with more than one -1 wildcard. Reshaping [(# 0 (SHAPE input_0)),2048,1,1] to [-1,-1].) out.log This was fixed by feature we released in 1.3. Closing this issue since it's resolved. @apbose Please open a new issue if you have any with the model.
gharchive/issue
2022-11-19T07:56:05
2025-04-01T04:35:36.906746
{ "authors": [ "Christina-Young-NVIDIA", "apbose", "peri044" ], "repo": "pytorch/TensorRT", "url": "https://github.com/pytorch/TensorRT/issues/1465", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
2516257776
Fix build error - error: function "torchao::marlin_24_gemm" has already been defined (previous definition at line 83) For CUDA_ARCH < 800 https://github.com/pytorch/ao/blob/3ac2ab8c9ffb0a624b48a51c105b0a780cafe483/torchao/csrc/cuda/sparse_marlin/marlin_kernel_nm.cu#L55-L57 we get multiple definition of marlin_24_gemm due to incorrect placement on endif. https://github.com/pytorch/ao/blob/3ac2ab8c9ffb0a624b48a51c105b0a780cafe483/torchao/csrc/cuda/sparse_marlin/marlin_kernel_nm.cu#L83-L92 https://github.com/pytorch/ao/blob/3ac2ab8c9ffb0a624b48a51c105b0a780cafe483/torchao/csrc/cuda/sparse_marlin/marlin_kernel_nm.cu#L1012-L1019 Fix - Move the endif to the appropriate location. mind just pasting the error you get before this change as well? otherwise lgtm Sure, I have updated the PR description. Thanks for the quick review :)
gharchive/pull-request
2024-09-10T12:07:22
2025-04-01T04:35:36.910019
{ "authors": [ "kshitij12345", "msaroufim" ], "repo": "pytorch/ao", "url": "https://github.com/pytorch/ao/pull/863", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
599331079
En-Fr result in scaling neural machine translation have compound spliting or not? As the name implies, the En-Fr BLEU in Scaling Neural Machine Translation have compound splitting or not? @edunov ? We include both results in the paper. We report 43.2 BLEU with compound splitting. The compound splitting is built-in to the preprocessing script). We also report 41.4 detokenized BLEU without compound splitting, using sacrebleu. We use the Moses detokenizer. bash scripts/compound_split_bleu.sh wmt16_gen.txt Can I use compound split bleu on En-Fr? No, don't use the compound_split_bleu.sh script for En-Fr. The underlying data already has compound splitting built-in. The BLEU score output from fairseq-generate (...) --remove-bpe can be reported directly as the compount-split BLEU score.
gharchive/issue
2020-04-14T06:35:36
2025-04-01T04:35:36.920885
{ "authors": [ "gaopengcuhk", "huihuifan", "myleott" ], "repo": "pytorch/fairseq", "url": "https://github.com/pytorch/fairseq/issues/2009", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
963119611
BPE and dictionary I run: fairseq-preprocess --tokenizer moses --bpe characters --source-lang src --target-lang tgt --trainpref $TEXT/train --testpref $TEXT/test --destdir data-bin2/data.toked.src-tgt With tokenizer and without, and with different tokenizers and bpes {characters/subword/..} but I always get the same vocabularies, with full words, shouldn't I see the words broken down in the vocabulary if I am using BPE? Env: fairseq master: PyTorch Version (e.g., 1.0) OS WSL/ Ubuntu /Windows doesn't make a difference How you installed fairseq pip/source doesn't make a difference What is bpe flag even doing? I put a breakpoint on "build_bpe" function and it is not hit, so even if I specify "bpe" in options, no bpe?
gharchive/issue
2021-08-07T00:44:50
2025-04-01T04:35:36.923198
{ "authors": [ "ImanHosseini" ], "repo": "pytorch/fairseq", "url": "https://github.com/pytorch/fairseq/issues/3767", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
537598288
Windows support? Currently I cannot install the package with MSVC toolset 14.11.25503 on Windows 10, Python 3.7 Anaconda. The command causing error is here. Hello @justanhduc - for now we don't test for explicit Windows coverage and don't maintain it. This will be rolled out incrementally.
gharchive/issue
2019-12-13T15:04:56
2025-04-01T04:35:36.939669
{ "authors": [ "cpuhrsch", "justanhduc" ], "repo": "pytorch/nestedtensor", "url": "https://github.com/pytorch/nestedtensor/issues/13", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1365359668
torchrec.inference.modules.quantize_embeddings does not seem to quantize weights. I tried to quantize my EBC in 3 different ways. None of them changed the dtype of the weights. Perhaps I'm misunderstanding when quantization happens? import torch import torch.quantization as quant import math import torchrec from torchrec.modules.embedding_configs import EmbeddingBagConfig from torchrec.modules.embedding_modules import EmbeddingBagCollection eb_configs = [ EmbeddingBagConfig( name=f"t_{feature_name}", embedding_dim=256, num_embeddings=1024, feature_names=[feature_name], ) for feature_idx, feature_name in enumerate( ['feature0', 'feature1',] ) ] ebc = EmbeddingBagCollection(tables=eb_configs, device='cuda') from torchrec.sparse.jagged_tensor import KeyedJaggedTensor features = KeyedJaggedTensor( keys=["feature0", "feature1"], values=torch.arange(1024).cuda(), lengths=torch.ones(1024).int().cuda(), ) ebc(features).values() print({key: value.dtype for key, value in ebc.state_dict().items()}) qebc = quant.quantize_dynamic(ebc, inplace=False) qebc(features).values() print({key: value.dtype for key, value in qebc.state_dict().items()}) from torchrec.inference.modules import quantize_embeddings tr_qebc = quantize_embeddings(ebc, dtype=torch.qint8, inplace=False) print({key: value.dtype for key, value in tr_qebc.state_dict().items()}) All the prints here output {'embedding_bags.t_feature0.weight': torch.float32, 'embedding_bags.t_feature1.weight': torch.float32} Hey @rllin, it seems like quantize_embeddings and quantize_dynamic only apply to submodules. With a wrapper module ebc = torch.nn.Sequential(ebc) I get: {'0.embedding_bags.t_feature0.weight': torch.uint8, '0.embedding_bags.t_feature1.weight': torch.uint8} Interesting, definitely not the best user experience, @s4ayub @zyan0 do you know why it's not applying quantization correctly to a top level EBC? The quantize_dynamic doc does seem to specify that it applies to submodules, not sure why they have this restriction: https://pytorch.org/docs/stable/generated/torch.quantization.quantize_dynamic.html @yifuwang thank this works!
gharchive/issue
2022-09-08T00:58:57
2025-04-01T04:35:37.141044
{ "authors": [ "YLGH", "rllin", "yifuwang" ], "repo": "pytorch/torchrec", "url": "https://github.com/pytorch/torchrec/issues/609", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
444812097
Fix link "getting to the venue" Missing hash in link deployed
gharchive/pull-request
2019-05-16T08:07:37
2025-04-01T04:35:37.263750
{ "authors": [ "Darkless012", "frenzymadness" ], "repo": "pyvec/cz.pycon.org-2019", "url": "https://github.com/pyvec/cz.pycon.org-2019/pull/184", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1894320371
Intermissions: talks and sponsors Integrace se sloty Zobrazují se 2 slidy s dalším talkem či streamovanou utility ze všech místností - uvažují se talky začínající v následující 1h 15m, abychom v sobotu dopoledne zbytečně nepromovali místnost __doc__, kde je něco až odpoledne. Talk se taky bude zobrazovat ještě 5 minut po plánovaném začátku, aby v případě zpoždění ještě chvilku zobrazovalo, co kde začíná. S aktuálním programem můžou nastat 3 situace: Následuje 1 talk: na obou slidech je stejný talk - bude to keynote/welcome/lightning, tak asi neva, když bude svítit o něco déle. Následují 3 talky: na prvním slidu bude 1 (ten z hlavní místnosti), na druhém 2. Následují 4 talky: zobrazení 2 + 2. Talky mají zatím úplně základní šablonu a layout. Sponzoři Při zadání levelu jde použít slicing, např: silver[:6] + silver[6:], zároveň těch levelů jde zadat více a zobrazí se vedle sebe, např. coffee,afterparty. Přidal jsem rendering loga a opět úplně základní layout, bude potřeba určitě ještě vyřešit velikosti atd. Další drobné změny Obrátil jsem ovládání pomocí šipek, doprava je nyní další slide, dříve to bylo obráceně. Do konzole v prohlížeči se vypisuje počet slidů a celková doba trvání cyklu + další debug informace (na něco je potřeba si explicitně zapnout level "verbose"). Doplnění: ještě jsem zapomněl zmínit, že GET parametrem time lze nasimulovat čas, pro který se mají talky zobrazit, např. http://localhost:8000/2023/intermissions/?time=2023-09-15T15:40 (ISO formát, lze použít i mezeru místo T)
gharchive/pull-request
2023-09-13T11:24:20
2025-04-01T04:35:37.267722
{ "authors": [ "jsmitka" ], "repo": "pyvec/cz.pycon.org", "url": "https://github.com/pyvec/cz.pycon.org/pull/56", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
631447693
Automatic Testing with CI I do not see any automatic unit testing setup in the project, what is the plan @pyvista/developers ? Nevermind, I just notice https://github.com/pyvista/pyvistaqt/blob/master/.ci/azure-pipelines.yml
gharchive/issue
2020-06-05T09:24:18
2025-04-01T04:35:37.284300
{ "authors": [ "GuillaumeFavelier" ], "repo": "pyvista/pyvistaqt", "url": "https://github.com/pyvista/pyvistaqt/issues/11", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1185276217
MNT: Readme This PR mainly updates some elements that were still using master as branch name. The rendered README.rst is available here: https://github.com/GuillaumeFavelier/pyvistaqt/tree/7c7283ebf40194cf486a064c26e35716d7135bc6 This is ready for review from my side @pyvista/pyvistaqt-admin One more commit for a small typo :smile: Thank you @akaszynski !
gharchive/pull-request
2022-03-29T18:50:05
2025-04-01T04:35:37.286443
{ "authors": [ "GuillaumeFavelier" ], "repo": "pyvista/pyvistaqt", "url": "https://github.com/pyvista/pyvistaqt/pull/156", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
369626756
Differing projection in annotator input Now that elements can be added as input to annotators, I would expect that each element could have its own projection. For example, I have a reservoir_df that is in UTM and a dam_df data frame that is in Plate Carree. I could feed these as data frames directly into the annotator, but to define the distinct projections, I converted them to gv objects each with their own projection. However, when I feed these into PolyAndPointAnnotator it doesn't honor them. If I give the annotator a crs, I think it overrules everything. If I don't give a crs and just give the gv objects (one as polys and one as points) it takes the first one and assumes the second is the same. Shouldn't it just honor the projection of the element? Closing.
gharchive/issue
2018-10-12T16:41:11
2025-04-01T04:35:37.288199
{ "authors": [ "kcpevey" ], "repo": "pyviz-topics/EarthSim", "url": "https://github.com/pyviz-topics/EarthSim/issues/247", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
391372597
Installing v0.3.0a7 with pip fails because of missing .version file setup.py/get_setup_version looks for a .version file (which doesn't exist) if the param package is not available. However, param does not seem to be a dependency of this project. This means I cannot currently pip install git+git://github.com/pyviz/hvplot.git@v0.3.0a7 without first installing param. For reference: I used pip v18.1 when I reported this issue. Thanks, I think we just need to add the pyproject.toml
gharchive/issue
2018-12-15T12:07:46
2025-04-01T04:35:37.290522
{ "authors": [ "lsorber", "philippjfr" ], "repo": "pyviz/hvplot", "url": "https://github.com/pyviz/hvplot/issues/128", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
588417650
feat(bigquery): add support of model for extract job Fixes [60] Merged.
gharchive/pull-request
2020-03-26T13:23:14
2025-04-01T04:35:37.304410
{ "authors": [ "HemangChothani" ], "repo": "q-logic/python-bigquery", "url": "https://github.com/q-logic/python-bigquery/pull/2", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1506656498
Dynamic struct encoding (#135) Taking over #135 adding abi encoder for the following data types dynamic struct static struct dynamic array of dynamic struct dynamic array of static struct static array of dynamic struct static array of static struct Codecov Report Merging #185 (1177a3d) into main (9e3e2af) will decrease coverage by 0.01%. The diff coverage is 100.00%. @@ Coverage Diff @@ ## main #185 +/- ## ========================================== - Coverage 99.75% 99.73% -0.02% ========================================== Files 70 70 Lines 4091 4229 +138 ========================================== + Hits 4081 4218 +137 - Misses 10 11 +1 Impacted Files Coverage Δ lib/eth/abi.rb 100.00% <100.00%> (ø) lib/eth/abi/type.rb 100.00% <100.00%> (ø) lib/eth/client.rb 100.00% <100.00%> (ø) lib/eth/contract/function.rb 100.00% <100.00%> (ø) lib/eth/contract/function_input.rb 100.00% <100.00%> (ø) spec/eth/abi/type_spec.rb 100.00% <100.00%> (ø) spec/eth/abi_spec.rb 100.00% <100.00%> (ø) spec/eth/contract/function_spec.rb 100.00% <100.00%> (ø) spec/eth/contract_spec.rb 100.00% <100.00%> (ø) lib/eth/client/ipc.rb 93.33% <0.00%> (-6.67%) :arrow_down: :mega: We’re building smart automated test selection to slash your CI/CD build times. Learn more
gharchive/pull-request
2022-12-21T17:18:37
2025-04-01T04:35:37.321663
{ "authors": [ "codecov-commenter", "q9f" ], "repo": "q9f/eth.rb", "url": "https://github.com/q9f/eth.rb/pull/185", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2175729087
Added basic todo app The app provides a simple boilerplate for CRUD (Create, Read, Update, Delete) operations. @ZeeshanMukhtar1 thanks, will review and merge this PR
gharchive/pull-request
2024-03-08T10:23:17
2025-04-01T04:35:37.485795
{ "authors": [ "ZeeshanMukhtar1", "qadir0108" ], "repo": "qadir0108/WebTechnologies", "url": "https://github.com/qadir0108/WebTechnologies/pull/11", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1990639412
🛑 nitter.us.projectsegfau.lt is down In 194e355, nitter.us.projectsegfau.lt (https://nitter.us.projectsegfau.lt) was down: HTTP code: 503 Response time: 392 ms Resolved: nitter.us.projectsegfau.lt is back up in fe6641b after 12 minutes.
gharchive/issue
2023-11-13T13:13:51
2025-04-01T04:35:37.489317
{ "authors": [ "qallen028" ], "repo": "qallen028/nitter-instances", "url": "https://github.com/qallen028/nitter-instances/issues/10120", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2014020160
🛑 nitter.projectsegfau.lt is down In 458f3d5, nitter.projectsegfau.lt (https://nitter.projectsegfau.lt) was down: HTTP code: 503 Response time: 363 ms Resolved: nitter.projectsegfau.lt is back up in e9d1328 after 11 minutes.
gharchive/issue
2023-11-28T09:34:35
2025-04-01T04:35:37.492575
{ "authors": [ "qallen028" ], "repo": "qallen028/nitter-instances", "url": "https://github.com/qallen028/nitter-instances/issues/11709", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1789116399
🛑 bird.trom.tf is down In 4f982c5, bird.trom.tf (https://bird.trom.tf) was down: HTTP code: 502 Response time: 452 ms Resolved: bird.trom.tf is back up in ad1c9d9.
gharchive/issue
2023-07-05T09:07:54
2025-04-01T04:35:37.495628
{ "authors": [ "qallen028" ], "repo": "qallen028/nitter-instances", "url": "https://github.com/qallen028/nitter-instances/issues/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1837948081
🛑 notabird.site is down In af3f41d, notabird.site (https://notabird.site) was down: HTTP code: 520 Response time: 72 ms Resolved: notabird.site is back up in 29dbead.
gharchive/issue
2023-08-05T21:37:47
2025-04-01T04:35:37.498052
{ "authors": [ "qallen028" ], "repo": "qallen028/nitter-instances", "url": "https://github.com/qallen028/nitter-instances/issues/2898", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1931672684
🛑 notabird.site is down In fb16801, notabird.site (https://notabird.site) was down: HTTP code: 520 Response time: 159 ms Resolved: notabird.site is back up in f24bcd2 after 1 hour, 8 minutes.
gharchive/issue
2023-10-08T06:45:33
2025-04-01T04:35:37.500739
{ "authors": [ "qallen028" ], "repo": "qallen028/nitter-instances", "url": "https://github.com/qallen028/nitter-instances/issues/7323", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1946749116
🛑 notabird.site is down In 15bb8eb, notabird.site (https://notabird.site) was down: HTTP code: 520 Response time: 190 ms Resolved: notabird.site is back up in 976fe12 after 5 minutes.
gharchive/issue
2023-10-17T07:35:53
2025-04-01T04:35:37.503017
{ "authors": [ "qallen028" ], "repo": "qallen028/nitter-instances", "url": "https://github.com/qallen028/nitter-instances/issues/7861", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
321765199
integrate qps-infra with smtp server by default it should use let's say public smtp creds or even smtp server w/o credentials. Ideally email notification should work right after install. Make sure to apply security if possible if any account creds are used won't fix as number of email notification via jenkins shold be minimal
gharchive/issue
2018-05-09T23:59:45
2025-04-01T04:35:37.504037
{ "authors": [ "vdelendik" ], "repo": "qaprosoft/qps-infra", "url": "https://github.com/qaprosoft/qps-infra/issues/16", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1643938043
Feature/bastion options Description Added new variable to be able to select Azure Bastion SKU (Basic and Standard), basic example fixed and updated. Module reconfigured with the necessary resources already specified within itself. Change overview (tick true): [x] This introduces backward incompatible changes [ ] This adds a new backward compatible Feature [x] This fixes a Bug Version information: [x] Current Version: major.minor.patch [ ] Next Version based on Semantic Versioning (see above): major.minor.patch How Has This Been Tested? [x] Apply of all examples was successful [ ] Test B Checklist: [x] I have run tests and documented them above [x] I have performed a self-review of my own code [ ] I have updated the documentation @JuanjoDoradoQB something went wrong here. According to Release v.1.0.4 this Pull-Request introduced breaking changes. According Semantic Versioning a breaking change should increase the major number. You correctly marked introducing new changes, but failed to update the version number. So for this PR it should have been this: - [ ] Current Version: `1.0.3` - [ ] Next Version based on [Semantic Versioning](https://semver.org/) (see above): `2.0.0` `` Please make sure to only create and approve PRs with correct version numbers. @QBYMKPI @QBY-MarkusMaring
gharchive/pull-request
2023-03-28T13:20:16
2025-04-01T04:35:37.514469
{ "authors": [ "JuanjoDoradoQB", "QBY-ThimoLimpert" ], "repo": "qbeyond/terraform-azurerm-azure-bastion", "url": "https://github.com/qbeyond/terraform-azurerm-azure-bastion/pull/2", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2141219513
Timer for bonus section of QBreader Would help to manage time when answering bonuses, similar to the timer on tus duplicate of #257
gharchive/issue
2024-02-18T22:38:23
2025-04-01T04:35:37.597570
{ "authors": [ "geoffrey-wu", "henrypanqb" ], "repo": "qbreader/website", "url": "https://github.com/qbreader/website/issues/260", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1665880593
chore!: Follow the naming standards in Open Controls According to our naming standards, package names, install names, and import names that include a product name shouldn't be preceded by "Q-CTRL". As "Open Controls" is listed as a product in the "Product" tab of the company website, the names in this package have to be updated to remove the initial "Q-CTRL" (except for the project name "Q-CTRL Python Open Controls"). Changes proposed in this pull request: Rename the package name to "Open Controls" in the README and documentation. Update the install name to open-controls. Update the import name to opencontrols. BREAKING CHANGE: Change import name to opencontrols. @tachikoma-li I guess we also need DevOp to rename this repo? These changes kind of assume no change in the name of the repo. Or, rather, they are agnostic about changes in the name of the repo. How does this work? Well, we follow the algorithm here and start with the repo name: repo name: python-open-controls To find the "Project name" we transform the slug into a sentence and add "Q-CTRL" to the front of it: Project name: Q-CTRL Python Open Controls Note that we're not touching the project name in this PR though, although we can change it later if we want to drop the "Python" part for example. Now, for the "Package name", we drop the language ("Python") and "Q-CTRL" if it is coming just before the name of a product ("Open Controls"). With this we get: Package name: Open Controls Once we have the package name, we slugify it to get the install name: Install name: open-controls and then remove the hyphens from the name to obtain the import name: Import name: opencontrols So this is basically just applying what's in the current naming conventions without changing the name of the repo. But if we're not happy with the results we can always change the naming conventions instead. @tachikoma-li I guess we also need DevOp to rename this repo? These changes kind of assume no change in the name of the repo. Or, rather, they are agnostic about changes in the name of the repo. Thanks! These all make sense. I guess one thing still a bit unclear to me is that it seems we begin with a repo name first and then define the project name. Is there any reason that we will keep using python-open-controls instead of open-controls? Or will that be done in another step? @tachikoma-li Is there any reason that we will keep using python-open-controls instead of open-controls? Or will that be done in another step? The latter, I was thinking of the change python-open-controls -> open-controls as a separate and independent step, which is why it is not covered here. I think we can do it after the renaming of the adapter packages (just in case renaming too many things at once causes some unexpected issue). Open Controls is no longer listed as a product in the products tab of the website, so these changes no longer apply. So I'm closing this
gharchive/pull-request
2023-04-13T07:32:23
2025-04-01T04:35:37.610523
{ "authors": [ "leoadec", "tachikoma-li" ], "repo": "qctrl/python-open-controls", "url": "https://github.com/qctrl/python-open-controls/pull/267", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1800946665
Specify the minimum Kubernetes version in pre-requisites Can you explicitly specify the minimum K8s version in README.md? https://github.com/qdrant/qdrant-helm#prerequisites @felix28 it's now in the readme 😃 https://github.com/qdrant/qdrant-helm/tree/main/charts/qdrant#prerequisites
gharchive/issue
2023-07-12T13:05:56
2025-04-01T04:35:37.613009
{ "authors": [ "cyril-corbon", "felix28" ], "repo": "qdrant/qdrant-helm", "url": "https://github.com/qdrant/qdrant-helm/issues/47", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
194624809
Ogc conformance testing This PR adds a quick tutorial how to setup the OGC confomance tests for WMS 1.3.0 and WMS 1.1.1 thanks @mhugent! @mhugent cool to see this. Do you have the QGIS project ? To run test ? @rldhont: the project does not involve styling at all. Just download the zip file with the data and load all the shapefiles and geotiffs into a project (the layer names should already be okay). Then you can run the tests against that project. @mhugent ok thanks!
gharchive/pull-request
2016-12-09T15:34:00
2025-04-01T04:35:37.621800
{ "authors": [ "mhugent", "rldhont", "yjacolin" ], "repo": "qgis/QGIS-Website", "url": "https://github.com/qgis/QGIS-Website/pull/403", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
139340266
warn if recommended artifact file extension isn't used Artifact and Visualization should warn when load or save is called with a file path that doesn't have the recommended file extension. This issue depends on #6. #106 added a warning when saving, does it also make sense to warn the user when loading with the wrong extension? @ebolyen and I talked about that, and it sounded like it made the greatest impact towards saving, as that can affect usage throughout the interfaces, because as of now the studio won't even acknowledge a non-.qza/v file, and the cli doesn't seem to care what you give it as long as it is loaded in without issue (e.g. a valid Artifact), (unless that itself is an issue?) but we weren't too sure on the warning for loading. I can definitely add it in though if it's something that should be there?
gharchive/issue
2016-03-08T17:25:59
2025-04-01T04:35:37.748386
{ "authors": [ "jairideout", "jakereps" ], "repo": "qiime2/qiime2", "url": "https://github.com/qiime2/qiime2/issues/7", "license": "bsd-3-clause", "license_type": "permissive", "license_source": "bigquery" }
2395050370
Are there any tips to improve the performance of GeoTransformer? When I apply geotransformer in my data,but get a bad result Maybe try https://github.com/qinzheng93/GeoTransformer/issues/116
gharchive/issue
2024-07-08T08:55:09
2025-04-01T04:35:37.749652
{ "authors": [ "cduch9", "zhuozhizhong" ], "repo": "qinzheng93/GeoTransformer", "url": "https://github.com/qinzheng93/GeoTransformer/issues/113", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
447034870
打开自动刷新,左侧db列表未刷新 操作系统OS Windows 10 应用版本 1.1.8 问题描述 其他客户端删除某一记录后,左侧列表中的key无法实时刷新,需要切换db再切回后才能刷新。 建议根据键值统计中数据,监测左侧key值列表。 如图 统计七条,显示8条 右侧自动刷新的初衷是只刷新右侧状态数据,并不会刷新左侧key列表 左侧列表可以通过点击搜索的放大镜实现刷新,或者按你说的切换db,后期也可以考虑再列表处加个刷新按钮
gharchive/issue
2019-05-22T09:42:02
2025-04-01T04:35:37.751916
{ "authors": [ "qishibo", "wilcool" ], "repo": "qishibo/AnotherRedisDesktopManager", "url": "https://github.com/qishibo/AnotherRedisDesktopManager/issues/30", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1077795473
Update 007_Reverse_Integer.py This is correct solution. Earlier solution was incorrect. Hi @mustafanw Thank you for contribution! Note that almost all Python solutions are based on Python2. These solutions may fail on Python3. I have checked 07 Python solution. It passed on Python2 and failed on Python3. So, this is not a solution problem but compatibility problem. In this case, I suggest add a new solution in Python 3 rather replace existing solution.
gharchive/pull-request
2021-12-12T12:32:58
2025-04-01T04:35:37.753713
{ "authors": [ "mustafanw", "qiyuangong" ], "repo": "qiyuangong/leetcode", "url": "https://github.com/qiyuangong/leetcode/pull/49", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
901555618
功能建议:希望添加上传跳过同名文件功能 原因是上传20000个小文件(10-600m大小)时,发现其中一个文件存在反复上传,不成功也不失败的状态。同名文件会覆盖的话也不好重新开始任务。 命令 参数是./BaiduPCS-Go upload --norapid --nosplit -l 4 /mnt/sdb1 . 反复上传不太清楚是什么状况,上传遇到同名文件默认会直接覆盖 下个版本会增加自动重命名选项 反复上传不太清楚是什么状况,上传遇到同名文件默认会直接覆盖 下个版本会增加自动重命名选项 不考虑跳过吗,这样也能用来同步文件 最后还是决定用跳过策略了
gharchive/issue
2021-05-26T00:35:56
2025-04-01T04:35:37.755737
{ "authors": [ "qjfoidnh", "zzzop" ], "repo": "qjfoidnh/BaiduPCS-Go", "url": "https://github.com/qjfoidnh/BaiduPCS-Go/issues/75", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
922958642
出大事了,原来没有同名文件提示! 上传的时候忘记改文件名,结果把原来的同名文件给覆盖了!40G的文件骤然消失,我太难了。 这个功能如果可以实现请务必增加。覆盖掉的文件也不是被删除了因为回收站里也没有,凭空消失无法找回啊! 心疼一波老哥......如果以前记录过文件的秒传信息可以通过秒传找回来,如果莫得。。。大概真就没了 软件的默认策略确实是覆盖,在实现逻辑上是通过百度网盘官方API的参数控制的,做出提示比较困难,但3.8.1可以通过--skip来修改重名策略为自动跳过,下个版本可能会增加自动重命名的策略选项。
gharchive/issue
2021-06-16T18:37:55
2025-04-01T04:35:37.757318
{ "authors": [ "mrdobejelly", "qjfoidnh" ], "repo": "qjfoidnh/BaiduPCS-Go", "url": "https://github.com/qjfoidnh/BaiduPCS-Go/issues/83", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
1338620115
Test Issue - cwyfv Test Body - rsjhndpbaa Test Comment - wxxbc
gharchive/issue
2022-08-15T07:05:38
2025-04-01T04:35:37.758278
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/10225", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1356268582
Test Issue - orvrs Test Body - zrgjggylgi Test Comment - nzoks
gharchive/issue
2022-08-30T20:29:34
2025-04-01T04:35:37.759056
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/11479", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1369153280
Test Issue - mqvir Test Body - bwkruljkgv Test Comment - hxxay
gharchive/issue
2022-09-12T01:40:51
2025-04-01T04:35:37.759759
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/12360", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1381443025
Test Issue - rfmei Test Body - sxmfczxlmw Test Comment - xjzko
gharchive/issue
2022-09-21T20:15:33
2025-04-01T04:35:37.760476
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/13218", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1400127823
Test Issue - gfutq Test Body - ritxduonac Test Comment - fzerj
gharchive/issue
2022-10-06T18:30:36
2025-04-01T04:35:37.761426
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/14416", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1403813740
Test Issue - mbwwk Test Body - reenijkkvo Test Comment - ktufl
gharchive/issue
2022-10-10T23:32:25
2025-04-01T04:35:37.762135
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/14655", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1429231571
Test Issue - srkru Test Body - bvdyhglphr Test Comment - wdzbc
gharchive/issue
2022-10-31T04:24:42
2025-04-01T04:35:37.762900
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/15906", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1218354167
Test Issue - mnztv Test Body - pqkftgyizb Test Comment - gxdgj
gharchive/issue
2022-04-28T08:03:07
2025-04-01T04:35:37.763591
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/1678", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1453848281
Test Issue - mwkdg Test Body - ottidsccjo Test Comment - aaiiv
gharchive/issue
2022-11-17T19:14:27
2025-04-01T04:35:37.764537
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/17079", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1220417537
Test Issue - rtouz Test Body - pkyscepbqt Test Comment - ebfol
gharchive/issue
2022-04-29T10:05:31
2025-04-01T04:35:37.765230
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/1796", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1499456338
Test Issue - aybrs Test Body - xumqbusmzy Test Comment - eussv
gharchive/issue
2022-12-16T02:19:57
2025-04-01T04:35:37.766036
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/18792", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1581469259
Test Issue - jmlwu Test Body - aiuqnzwhjm Test Comment - shftv
gharchive/issue
2023-02-13T00:29:26
2025-04-01T04:35:37.766844
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/21851", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1654425743
Test Issue - qhcni Test Body - zuufvmcluj Test Comment - lbvwd
gharchive/issue
2023-04-04T19:06:00
2025-04-01T04:35:37.767551
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/24084", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1672046782
Test Issue - npnnz Test Body - rdxsnvcmow Test Comment - kcleo
gharchive/issue
2023-04-17T22:14:32
2025-04-01T04:35:37.768368
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/24634", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1798746102
Test Issue - vdrrz Test Body - eipeirogxs Test Comment - vwfdb
gharchive/issue
2023-07-11T11:42:10
2025-04-01T04:35:37.769164
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/28319", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1815119415
Test Issue - fzzui Test Body - ahbhrppgbq Test Comment - unwqj
gharchive/issue
2023-07-21T04:06:33
2025-04-01T04:35:37.769871
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/28791", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1895681820
Test Issue - zmucj Test Body - axlnjsjahe Test Comment - zapkq
gharchive/issue
2023-09-14T05:06:24
2025-04-01T04:35:37.770873
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/31609", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1944687553
Test Issue - thgby Test Body - kligyczavw Test Comment - vpszw
gharchive/issue
2023-10-16T08:41:44
2025-04-01T04:35:37.771666
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/33190", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1951253494
Test Issue - bwsgn Test Body - wyebayspja Test Comment - yprfx
gharchive/issue
2023-10-19T05:14:24
2025-04-01T04:35:37.772388
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/33361", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1959483133
Test Issue - sdxvh Test Body - dywrfrxuxk Test Comment - wmexj
gharchive/issue
2023-10-24T15:06:56
2025-04-01T04:35:37.773183
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/33631", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2027270524
Test Issue - aifta Test Body - pqhcfmzblh Test Comment - aetsn
gharchive/issue
2023-12-05T22:51:56
2025-04-01T04:35:37.774067
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/35814", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2071661301
Test Issue - eggdg Test Body - vtnjnpwsei Test Comment - qrvsm
gharchive/issue
2024-01-09T05:35:42
2025-04-01T04:35:37.774764
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/37565", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2105176927
Test Issue - kbtiz Test Body - nqebttvbuc Test Comment - zjhwg
gharchive/issue
2024-01-29T10:34:45
2025-04-01T04:35:37.775511
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/38401", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2219445905
Test Issue - dsnhh Test Body - ajpldgfurg Test Comment - idcvg
gharchive/issue
2024-04-02T02:39:50
2025-04-01T04:35:37.776284
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/42264", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2372142800
Test Issue - lbdcx Test Body - tendvhbcvj Test Comment - saufo
gharchive/issue
2024-06-25T09:14:48
2025-04-01T04:35:37.776998
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/47322", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2403712445
Test Issue - rpvtx Test Body - hobkdpfufn Test Comment - vsetv
gharchive/issue
2024-07-11T17:38:26
2025-04-01T04:35:37.777756
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/48357", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2413719960
Test Issue - niupc Test Body - mdqsiwyazf Test Comment - pepyg
gharchive/issue
2024-07-17T14:06:09
2025-04-01T04:35:37.778539
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/48678", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1266677223
Test Issue - vrvpi Test Body - dlwlpdoteg Test Comment - qluou
gharchive/issue
2022-06-09T21:25:42
2025-04-01T04:35:37.779290
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/5061", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2529503499
Test Issue - desmo Test Body - pouozvcahr Test Comment - tznyp
gharchive/issue
2024-09-16T21:08:19
2025-04-01T04:35:37.780263
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/52310", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2555340077
Test Issue - nyueh Test Body - gfvvrusdxt Test Comment - fysui
gharchive/issue
2024-09-30T01:35:10
2025-04-01T04:35:37.780967
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/53088", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2563906377
Test Issue - gnlyi Test Body - mjdzwbwfjt Test Comment - mmdjd
gharchive/issue
2024-10-03T12:16:45
2025-04-01T04:35:37.781796
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/53381", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2595187383
Test Issue - tyhxu Test Body - otrouuotuq Test Comment - qppaz
gharchive/issue
2024-10-17T16:15:57
2025-04-01T04:35:37.782605
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/54233", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2626054025
Test Issue - bmdbt Test Body - hiewqvsiyh Test Comment - bimcc
gharchive/issue
2024-10-31T06:15:59
2025-04-01T04:35:37.783316
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/55035", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
2697548076
Test Issue - ylmen Test Body - stydtsjkdj Test Comment - mwwad
gharchive/issue
2024-11-27T07:59:05
2025-04-01T04:35:37.784101
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/56197", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1308868509
Test Issue - ivxwz Test Body - aarefscjgi Test Comment - izato
gharchive/issue
2022-07-19T01:41:34
2025-04-01T04:35:37.784863
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/8075", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
1318880717
Test Issue - dazpb Test Body - sslqfiwona Test Comment - eadtw
gharchive/issue
2022-07-27T00:23:19
2025-04-01T04:35:37.785604
{ "authors": [ "qlikqaa" ], "repo": "qlikqaa/dummyrepo", "url": "https://github.com/qlikqaa/dummyrepo/issues/8712", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
326726243
How to stop the daemon ? I wanted to know how to stop the daemon... I found that's kill 12998
gharchive/issue
2018-05-26T09:59:30
2025-04-01T04:35:37.840727
{ "authors": [ "dimensi0n" ], "repo": "qodrorid/godaemon", "url": "https://github.com/qodrorid/godaemon/issues/1", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
191670356
CSF Configurations button non visible Hello, After updating CSF and authentic theme to the latest versions I realised that the CSF fw configuration button was not visible at all. I found that I had to DISABLE the custom style within the CSF configuration (via command line >> STYLE_CUSTOM = "0") as this was enabled (not by me). After that it worked, however my question here is... will this affect the styling imposed by your theme anyhow..or it is OK to leave the above setting disabled? Thank you. Configuration button defaults to the default place in Webmin - table header. Look up at the table's left. You will find it there.
gharchive/issue
2016-11-25T10:28:05
2025-04-01T04:35:37.845862
{ "authors": [ "netizenuk", "qooob" ], "repo": "qooob/authentic-theme", "url": "https://github.com/qooob/authentic-theme/issues/588", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
200798514
Theme: Virtualmin password recovery When using Virtualmin Password Recovery and Version 18.30-18.32 the login url for Version 18.30-18.32 is missing at the main page. Works with other themes. Hi, It's not missing. Have a look at the screenshot. I know about that text link. We decided to move it to the button. Maybe the text should be a bit more clear. Missing at 10 different servers when using authentic-theme and centos 7 Could you please click Reset orange button!? To be honest, it also is not clear to me that the Reset button is (or should be) a reset password button. It only seems to empty the Username and Password fields. What would be the good name? Password reset? If that is what it does, yes. But currently it only seems to clear the Username en Password fields on Firefox and Chrome. Didn't test any other. Server is running latest Webmin/Virtualmin running on Debian 64bit. Missing it? Do you have Virtualmin password recovery installed? As far as I remember it requires it to be installed in order to work. I never heard of the module, so I don't have it installed :p I also cannot find it inside Webmin/Virtualmin as an unused module. The only way to get it seems to be from here http://www.webmin.com/cgi-bin/search_third.cgi?category=Virtualmin. Maybe the theme shouldn't show the Reset (password) button if the required module is not installed, since the function of the button without this module can be confusing . I assume this lead to @ai-hosting opening this issue. Reset button is not needed I agree. I will see what I can do here. if Virtualmin Password Recovery is installed RESET will take you to the pass recovery url. Only problem is that is doens't support real password reset. If virtualmin allows plain text password storage it will send the old password by email. A smart admin will store the password in a hashed form. Making the Virtualmin Password Recovery unusable. @ai-hosting Even though I agree, that is not the theme issue. Report it to Webmin repo. What are we talking here? I'm not sure. I will remove reset button as it's useless. But if you have VM password recovery does it work the same way for you as in old themes? .. I will call the new button for Password recovery as such. Oh, yeah. Don't bother. I see what you mean. I will make a quick fix for you to check. Please have a try this patch.
gharchive/issue
2017-01-14T11:43:38
2025-04-01T04:35:37.854385
{ "authors": [ "ai-hosting", "jobhh", "qooob" ], "repo": "qooob/authentic-theme", "url": "https://github.com/qooob/authentic-theme/issues/646", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
595335717
Fix rotation bobbing issue Hey, I fixed a little bug where the hand of the player would bob if the portal rotation and the destination rotation are not equal. This was fixed by setting the renderYaw and renderPitch, as well as the lastRenderYaw and lastRenderPitch to the yaw of the player to prevent the renderer from interpolating incorrectly. Thanks
gharchive/pull-request
2020-04-06T18:33:40
2025-04-01T04:35:37.863378
{ "authors": [ "benfah", "qouteall" ], "repo": "qouteall/ImmersivePortalsMod", "url": "https://github.com/qouteall/ImmersivePortalsMod/pull/139", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2264493100
build: Tell cmake to set 'rpath' intelligently Before this commit, when QPDF's products were installed in an unorthodox location, it was possible for one of the executable products not to be able to find the necessary library product (i.e., the necessary libqpdf); that is, an executable didn't inform the dynamic linker where to look. For example: $ ldd /path/to/installed/files/bin/qpdf linux-gate.so.1 (0xb7ee6000) libqpdf.so.29 => not found [...] Now, because of this commit, an executable's rpath setting (or equivalent) is set upon installation, thereby providing the required search directory: $ ldd /path/to/installed/files/bin/qpdf linux-gate.so.1 (0xb7ee6000) libqpdf.so.29 => /path/to/installed/files/lib/libqpdf.so.29 (0xb7bc3000) [...] There is some intelligence behind whether rpath is set at all: If CMAKE_INSTALL_RPATH is set, then it is used; this allows the user to set a single, overriding value, and it allows the AppImage build script to continue working in exactly the same way that it has been doing so far. Otherwise, the base-case rpath is determined by the following:"${CMAKE_INSTALL_FULL_LIBDIR}" If that path is a standard location, then the base-case rpath is the empty string (meaning, in the base case, no rpath will be set); otherwise, the base-case rpath is set to that path. The base-case rpath is used to set various targets' INSTALL_RPATH property. This is done explicitly, so as to avoid the setting of an rpath unnecessarily on some target; for instance, as of this commit, only some of the executable targets have this property set, but the libqpdf target does not. In the long run, having such fine-grained control will likely be the best policy. In addition, cmake has been instructed to add to each target's rpath any directories that it thinks might contain potential unorthodox dependencies from outside the project; most of the time, there will be no such dependencies, and so nothing will be added. However, this is a simple way to account for unforeseen needs. Of course, this will only help on a system that supports an rpath feature known to cmake; for example, Windows has no such feature, and so all of this will presumably be ignored when building under that system. qpdf's handling of rpath has been problematic in various ways. I think you're on a good path here, but it's important to ensure that rpath is not set for distributions that install in temporary locations for packages. I think this is fine because they typically use DESTDIR. I will probably want to add somthing under pkg-test to ensure that the desired rpath behavior is present and then hope this doesn't cause a regression on any of the older rpath-related issues.
gharchive/pull-request
2024-04-25T21:04:30
2025-04-01T04:35:37.982243
{ "authors": [ "jberkenbilt", "mfwitten" ], "repo": "qpdf/qpdf", "url": "https://github.com/qpdf/qpdf/pull/1192", "license": "apache-2.0", "license_type": "permissive", "license_source": "bigquery" }
2042052626
There may be an issue with the I2C code I modified the print_ I2c_ Scan() section code ,I modified the I2C PINS and now I can find the I2C device. So now I can confirm that the software is not working properly void print_i2c_scan() { //Serial.printf("I2C: Scanning ...\n"); // byte count = 0; //i2c->begin(); //for (byte i = 8; i < 120; i++) { // i2c->beginTransmission(i); // Begin I2C transmission Address (i) // if (i2c->endTransmission() == 0) { // Receive 0 = success (ACK response) // Serial.printf("I2C: Found address: 0x%02X (%d)\n",i,i); // count++; // } //} //Serial.printf("I2C: Found %d device(s)\n", count); byte error, address; int nDevices; Serial.println("Scanning..."); nDevices = 0; for(address = 1; address < 127; address++ ) { // The i2c_scanner uses the return value of // the Write.endTransmisstion to see if // a device did acknowledge to the address. WIRE.beginTransmission(address); error = WIRE.endTransmission(); if (error == 0) { Serial.print("I2C device found at address 0x"); if (address<16) Serial.print("0"); Serial.print(address,HEX); Serial.println(" !"); nDevices++; } else if (error==4) { Serial.print("Unknown error at address 0x"); if (address<16) Serial.print("0"); Serial.println(address,HEX); } } if (nDevices == 0) Serial.println("No I2C devices found\n"); else Serial.println("done\n"); } madflight starting 5 ... madflight starting 4 ... madflight starting 3 ... madflight starting 2 ... madflight starting 1 ... USE_HW_ESP32 I2C: SDA=21 SCL=22 SPI: MOSI=23 MISO=19 SCLK=18 CS=15 Scanning... I2C device found at address 0x68 ! done USE_RX_PPM pin=35 USE_IMU_MPU6050_I2C E (4065) ledc: requested frequency and duty resolution can not be achieved, try reducing freq_hz or duty_resolution. div_param=48 E (4066) ledc: requested frequency and duty resolution can not be achieved, try reducing freq_hz or duty_resolution. div_param=97 E (4077) ledc: requested frequency and duty resolution can not be achieved, try reducing freq_hz or duty_resolution. div_param=195 E (4089) ledc: requested frequency and duty resolution can not be achieved, try reducing freq_hz or duty_resolution. div_param=48 E (4100) ledc: requested frequency and duty resolution can not be achieved, try reducing freq_hz or duty_resolution. div_param=97 E (4111) ledc: requested frequency and duty resolution can not be achieved, try reducing freq_hz or duty_resolution. div_param=195 E (4123) ledc: requested frequency and duty resolution can not be achieved, try reducing freq_hz or duty_resolution. div_param=48 E (4134) ledc: requested frequency and duty resolution can not be achieved, try reducing freq_hz or duty_resolution. div_param=97 E (4145) ledc: requested frequency and duty resolution can not be achieved, try reducing freq_hz or duty_resolution. div_param=195 E (4157) ledc: requested frequency and duty resolution can not be achieved, try reducing freq_hz or duty_resolution. div_param=48 E (4168) ledc: requested frequency and duty resolution can not be achieved, try reducing freq_hz or duty_resolution. div_param=97 E (4180) ledc: requested frequency and duty resolution can not be achieved, try reducing freq_hz or duty_resolution. div_param=195 float AccErrorX = +0.000061; float AccErrorY = +0.000061; float AccErrorZ = -1.000061; float GyroErrorX = -0.007634; float GyroErrorY = -0.007634; float GyroErrorZ = +0.007634; Paste these values in user specified variables section and comment out calculate_IMU_error() in void setup. ahrs_Setup() No Magnetometer ahrs_Setup() Estimated yaw:-0.00 gx:+0.00 gy:+0.00 gz:-0.00 ax:+0.00 ay:+0.00 az:-0.00 mx:+0.00 my:+0.00 mz:+0.00 roll:+0.0 pitch:-0.0 yaw:-0.0 yaw_mag:-0.0 imu_err_cnt:0 gx:+0.00 gy:+0.00 gz:-0.00 ax:+0.00 ay:+0.00 az:-0.00 mx:+0.00 I used the code here https://learn.adafruit.com/scanning-i2c-addresses/arduino Hi, thanks for trying this out. As I understand it, you changed the SDA pin from 23 to 21 in hw_ESP32.h. But did you also change the HW_PIN_SPI_MOSI pin from 21 to 23??? Otherwise you have pin 21 defined twice, that does not work. You should have something like this //I2C: const int HW_PIN_I2C_SDA = 21; //<<<<<<<<<<<<<<<<<< was 23 const int HW_PIN_I2C_SCL = 22; //default: Wire 22 typedef SoftWire HW_WIRETYPE; //define the class to use for I2C //TwoWire *i2c = &Wire; //&Wire or &Wire1 - when using <Wire.h> HW_WIRETYPE *i2c = new HW_WIRETYPE(); //create a ESP32_SoftWire instance //SPI: const int HW_PIN_SPI_MOSI = 23; //<<<<<<<<<<<<<<<<<<<<<< was 21 const int HW_PIN_SPI_MISO = 36; //VP defaults: VSPI 19, HSPI 12 const int HW_PIN_SPI_SCLK = 19; // defaults: VSPI 18, HSPI 14 const int HW_PIN_SPI_CS = 18; // defaults: VSPI 5, HSPI 15 SPIClass *spi = new SPIClass(HSPI); // VSPI or HSPI(default) Maybe try a fresh git clone, I've been making quite a few updates the last couple of days. Here is the output I'm getting on an ESP32 with MPU9150 (= MPU6050 + magnetometer) connected. USE_HW_ESP32 I2C: SDA=21 SCL=22 SPI: MOSI=23 MISO=36 SCLK=19 CS=18 I2C: Scanning ... I2C: Found address: 0x69 (105) I2C: Found address: 0x77 (119) I2C: Found 2 device(s) USE_RCIN_CRSF USE_IMU_MPU6050_I2C .... CH1:1500 rcin_roll:+0.00 gx:-7.00 ax:-0.00 mx:+0.00 ahrs_roll:-0.1 roll_PID:+0.004 m1%:0 loop_rt:375 CH1:1500 rcin_roll:+0.00 gx:-7.00 ax:-0.00 mx:+0.00 ahrs_roll:-0.2 roll_PID:+0.004 m1%:0 loop_rt:354 CH1:1500 rcin_roll:+0.00 gx:-7.21 ax:-0.01 mx:+0.00 ahrs_roll:-0.2 roll_PID:+0.004 m1%:0 loop_rt:342 Thank you! Pin 21 is only defined once ,MPU9150 (=MPU6500+magnetometer), so MPU9250 and MPU6050 are completely different,I will test MPU6500 immediately //I2C: const int i2c_SDA_PIN = 21; //default: Wire 21 const int i2c_SCL_PIN = 22; //default: Wire 22 typedef SoftWire HW_WIRETYPE; //define the class to use for I2C //TwoWire *i2c = &Wire; //&Wire or &Wire1 - when using <Wire.h> HW_WIRETYPE *i2c = new HW_WIRETYPE(); //create a ESP32_SoftWire instance //SPI: const int spi_MOSI_PIN = 23; // default: VSPI 23, HSPI 13 const int spi_MISO_PIN = 19; //VP default: VSPI 19, HSPI 12 const int spi_SCLK_PIN = 18; // default: VSPI 18, HSPI 14 const int spi_CS_PIN = 15; // default: VSPI 5, HSPI 15 SPIClass *spi = new SPIClass(VSPI); // VSPI(default) or HSPI I tested using MPU6500, but your code still cannot recognize the device I am using the ARDUINO1.8.16IDE and I have used the development board linked below NodeMCU ESP32 S Dev Kit https://www.az-delivery.de/products/nodemcu-esp-32s-kit?shpxid=f07075dc-6a71-4466-9b0e-27f241a6759f But I use RP2040+MPU6500, and your code can identify the device madflight starting 5 ... madflight starting 4 ... madflight starting 3 ... madflight starting 2 ... madflight starting 1 ... USE_HW_RP2040 I2C: SDA=20 SCL=21 SPI: MOSI=19 MISO=16 SCLK=18 CS=17 I2C: Scanning ... I2C: Found address: 0x68 (104) I2C: Found 1 device(s) USE_RX_PPM pin=1 USE_IMU_MPU9250_I2C I used the code you updated today, RP2040 can recognize MPU6050 and MPU6500, but ESP32 cannot recognize MPU6050 and MPU6500. RP2040 can only recognize that the device has no XYZ data output when connected to MPU6050. RP2040 can recognize that the device has XYZ data output when connected to MPU6500. RP2040 connected to MPU6050 madflight starting 3 ... madflight starting 2 ... madflight starting 1 ... USE_HW_RP2040 HW_PIN_LED=25 HW_PIN_RCIN_RX=1 TX=0 HW_PIN_IMU_INT=22 HW_PIN_I2C_SDA=20 SCL=21 HW_PIN_SPI_MOSI=19 MISO=16 SCLK=18 CS=17 HW_PIN_OUT[14]=2,3,4,5,6,7,8,9,10,11,12,13,14,15 I2C: Scanning ... I2C: Found address: 0x68 (104) I2C: Found 1 device(s) USE_RCIN_CRSF USE_IMU_MPU6050_I2C RP2040 connected to MPU6500. madflight starting 5 ... madflight starting 4 ... madflight starting 3 ... madflight starting 2 ... madflight starting 1 ... USE_HW_RP2040 HW_PIN_LED=25 HW_PIN_RCIN_RX=1 TX=0 HW_PIN_IMU_INT=22 HW_PIN_I2C_SDA=20 SCL=21 HW_PIN_SPI_MOSI=19 MISO=16 SCLK=18 CS=17 HW_PIN_OUT[14]=2,3,4,5,6,7,8,9,10,11,12,13,14,15 I2C: Scanning ... I2C: Found address: 0x68 (104) I2C: Found 1 device(s) USE_RCIN_CRSF USE_IMU_MPU9250_I2C IMU init failed rv=-1112. Retrying... USE_IMU_MPU9250_I2C IMU init failed rv=-1000. Retrying... USE_IMU_MPU9250_I2C IMU init failed rv=-1000. Retrying... USE_IMU_MPU9250_I2C IMU init failed rv=-1000. Retrying... USE_IMU_MPU9250_I2C IMU init failed rv=-1000. Retrying... USE_IMU_MPU9250_I2C IMU init failed rv=-1000. Retrying... USE_IMU_MPU9250_I2C IMU init failed rv=-1000. Retrying... USE_IMU_MPU9250_I2C IMU init failed rv=-1000. Retrying... USE_IMU_MPU9250_I2C IMU init failed rv=-1000. Retrying... USE_IMU_MPU9250_I2C IMU init failed rv=-1000. Retrying... ahrs_Setup() Estimated yaw:-180.00 CH1:1459 rcin_roll:+0.00 gx:-95.15 ax:-0.38 mx:+0.00 ahrs_roll:-0.1 roll_PID:+0.048 m1%:0 loop_rt:891 gx:-95.15 gy:-107.25 gz:+125.13 ax:-0.38 ay:+0.13 az:-0.18 mx:+0.00 my:+0.00 mz:+0.00 roll:-0.1 pitch:-0.1 yaw:-179.9 yaw_mag:-0.0 CH1:1500 rcin_roll:+0.00 gx:+25.71 ax:-1.29 mx:+0.00 ahrs_roll:-0.5 roll_PID:-0.012 m1%:0 loop_rt:564 gx:+25.71 gy:-8.97 gz:-8.10 ax:-1.29 ay:-1.32 az:+0.96 mx:+0.00 my:+0.30 mz:+0.00 roll:-0.5 pitch:-0.3 yaw:-179.3 yaw_mag:-90.0 CH1:1500 rcin_roll:+0.00 gx:+25.91 ax:-1.29 mx:+0.00 ahrs_roll:-0.4 roll_PID:-0.012 m1%:0 loop_rt:646 gx:+25.91 gy:-8.81 gz:-8.32 ax:-1.29 ay:-1.33 az:+0.97 mx:+0.00 my:+0.30 mz:+0.00 roll:-0.4 pitch:+0.0 yaw:-179.4 yaw_mag:-90.0 CH1:1500 rcin_roll:+0.00 gx:+25.91 ax:-1.29 mx:+0.00 ahrs_roll:-0.2 roll_PID:-0.012 m1%:0 loop_rt:663 gx:+25.91 gy:-9.04 gz:-8.11 ax:-1.29 ay:-1.32 az:+0.97 mx:+0.00 my:+0.30 mz:+0.00 roll:-0.2 pitch:-0.2 yaw:-178.8 yaw_mag:-90.0 CH1:1500 rcin_roll:+0.00 gx:+25.91 ax:-1.29 mx:+0.00 ahrs_roll:-0.1 roll_PID:-0.013 m1%:0 loop_rt:673 gx:+25.91 gy:-8.81 gz:-8.32 ax:-1.29 ay:-1.33 az:+0.97 mx:+0.00 my:+0.30 mz:+0.00 roll:-0.1 pitch:+0.1 yaw:-178. I've updated imu.h, RP2040 connected to MPU6500 should work now. "NodeMCU ESP32 S Dev Kit" - I'm not 100% sure this is a ESP32 or a ESP32-Sx variant, or some sort of fake. I've never seen the ESP32S module on the picture. Anyway, I've added #define USE_ESP32_SOFTWIRE in hw_ESP32.h, comment this line to use the regular Wire.h library. Maybe it helps. Please give it a try, and let me know. comment this line# define USE_ ESP32_ SOFTWIRE to use the regular Wire. h library.Will cause ESP32 to fail compilation RP2040 connected to MPU6500, Still only identifying the device, MPU6050 cannot output correct data I have purchased a new MPU6500 module and will continue testing it tomorrow Will your code support WING in the future? Will the barometer and GPS be added to provide automatic home function Thank you! Pushed a new version today with some barometers. Still fighting with I2C on the ESP32, added a note in hw_ESP32.h Did not test on RP2040 today, hope it still works ;-) Will your code support WING in the future? Will the barometer and GPS be added to provide automatic home function If possible, it would be better to support BMI 323 and ICM42688, as both BF and INAV now support these sensors. adding an SD card holder can record some data. If WIFI is utilized, the PID can be modified using web mode These are certainly options. I'm working on barometers now, then maybe GPS. The rest probably won't happen (not enough time to work on this...) But if you feel like, please go ahead and add the new functions/sensors and send me a pull request. I am planning to design a 36X36 FC hardware Sounds good! Actually RP2040 can do 6 hardware serial ports with the PIO see here Thank you again! Attempting to connect to MPU6500 through I2C mode using RP2040 failed again, but when I connect to MPU6500 through SPI mode using RP2040, the measurement data of MPU6500 can be displayed.I have abandoned I2C mode. I should enter the hardware design phase now, using RP2040+ELRS (SX1280+ESP8285)+MPU6500(or MPU9250,Actually, I prefer BMI 088 more )+BMP280(Actually, I prefer DPS310 more) +5VBEC Can you give some advice on hardware design? Definitely use SPI for the IMU: faster and more stable. I would not bother to include ELRS on the board, but provide a 4 pin connector to connect an external ELRS module. Rather include a current shunt and voltage sense resistor divider (see new BAT_ADC pin on my pinout) or use an INA226. You can use more modern IMU/BARO chips, but you will need to write drivers for them as I'm focusing on easily available sensor breakout boards at the moment. Options like LCD or SDCARD will probably need SPI, which is already used for IMU or you need to sacrifice 4 pins for a second SPI. Looking forward to see your design!
gharchive/issue
2023-12-14T16:32:35
2025-04-01T04:35:38.021668
{ "authors": [ "01Space", "qqqlab" ], "repo": "qqqlab/madflight", "url": "https://github.com/qqqlab/madflight/issues/3", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }
127139651
doesn't work I was going to test it for some advanced features, but it fails on the most basic stuff: select * from test outputs: elect * from test Hi there. Thank you for reporting. It seems to behave strangely with CRLF files (Windows system' end-of-line character). I'll look into it. Just pushed version 1.0.4 with support for Windows files. npm update -g sql-uglifier Please check if it is working properly now. The bug I described was fixed, but the next issue I saw - it doesn't remove single-line --comments. Anyway, since I needed a reliable minification specifically for PostgreSQL scripts, I wrote one of my own yesterday - pg-minify. Thank you for your support!
gharchive/issue
2016-01-18T01:45:21
2025-04-01T04:35:38.026341
{ "authors": [ "qrawl", "vitaly-t" ], "repo": "qrawl/sql-uglifier", "url": "https://github.com/qrawl/sql-uglifier/issues/1", "license": "mit", "license_type": "permissive", "license_source": "bigquery" }
1761671096
[Feature Request]Enforce submission despite test case mismatch I whould be great to have the ability to force submissions even if there are test case discrepancies. Like in the following problem, when multiple output results are allowed, we will not be able to submit with cargo-compete. https://atcoder.jp/contests/typical90/tasks/typical90_z --no-test Do not test before submitting
gharchive/issue
2023-06-17T06:23:02
2025-04-01T04:35:38.034693
{ "authors": [ "nayo0513" ], "repo": "qryxip/cargo-compete", "url": "https://github.com/qryxip/cargo-compete/issues/201", "license": "Apache-2.0", "license_type": "permissive", "license_source": "github-api" }
2194416931
Qcodes driver fix qua_program replaced by a field numpy import fix Hey @yomach, this relates to this issue, is it what you had in mind? I'm not sure. Since we didn't reproduce the issue locally I don't know if this fixed it. We can try to ask the author of that issue in the ticket. I'll ping him Here we are waiting for the response from the guy who asked for the PR. I won't merge until he approves and I just rebased it on the new version, sorry for the useless notification... Envoyé à partir de Outlook pour Androidhttps://aka.ms/AAb9ysg From: Yoav Romach @.> Sent: Thursday, April 18, 2024 7:08:10 PM To: qua-platform/py-qua-tools @.> Cc: Theo Laudat @.>; Author @.> Subject: Re: [qua-platform/py-qua-tools] Qcodes driver fix (PR #190) @yomach commented on this pull request. Changelog? — Reply to this email directly, view it on GitHubhttps://github.com/qua-platform/py-qua-tools/pull/190#pullrequestreview-2009450646, or unsubscribehttps://github.com/notifications/unsubscribe-auth/AXR3HVT7QHPXUWQMO3KXHE3Y574XVAVCNFSM6AAAAABE5EIYI2VHI2DSMVQWIX3LMV43YUDVNRWFEZLROVSXG5CSMV3GSZLXHMZDAMBZGQ2TANRUGY. You are receiving this because you authored the thread.Message ID: @.***>
gharchive/pull-request
2024-03-19T09:11:30
2025-04-01T04:35:38.057547
{ "authors": [ "TheoLaudatQM", "yomach" ], "repo": "qua-platform/py-qua-tools", "url": "https://github.com/qua-platform/py-qua-tools/pull/190", "license": "BSD-3-Clause", "license_type": "permissive", "license_source": "github-api" }
1605035516
🛑 otakusan is down In 7a24433, otakusan (https://otakusan.net/LightNovel) was down: HTTP code: 0 Response time: 0 ms Resolved: otakusan is back up in 916b119.
gharchive/issue
2023-03-01T14:02:38
2025-04-01T04:35:38.070454
{ "authors": [ "quanhieu" ], "repo": "quanhieu/alive_up", "url": "https://github.com/quanhieu/alive_up/issues/1402", "license": "MIT", "license_type": "permissive", "license_source": "github-api" }