id stringlengths 4 10 | text stringlengths 4 2.14M | source stringclasses 2
values | created timestamp[s]date 2001-05-16 21:05:09 2025-01-01 03:38:30 | added stringdate 2025-04-01 04:05:38 2025-04-01 07:14:06 | metadata dict |
|---|---|---|---|---|---|
1350443897 | Allow for custom "shops" table for automatic migrations
laravel-shopify allowing custom auth guard in their latest version. But it not support to allowing migration with custom table for shops in automatic migration.
This PR is allowing developer's to set their own shops table in config to execute migration with given name.
@osiset, I have already rewritten the code based on your change request. Please review the PR again. Thank you
@nahid I agree. Go ahead with it :)
@osiset, I have pushed the following changes, Please review again
Solid. I'll approve letting Actions run and we'll see what happens. Probably will need test added to Until tests to ensure coverage is there.
Solid. I'll approve letting Actions run and we'll see what happens. Probably will need test added to Until tests to ensure coverage is there.
Maybe tests/Storage/Queries/ChargeTest.php file needs to be updated to code coverage. Please let me know if there is anything I can do. Thank you
@nahid Tests pass, coverage didnt change as well, so thats OK.
Linting failed though, please see the Action for details.
@nahid Tests pass, coverage didnt change as well, so thats OK.
Linting failed though, please see the Action for details.
@osiset, I just pushed another commit with fixes, please check
@nahid Still is failing the lint.
Locally, you can vendor/bin/php-cs-fixer fix --diff --dry-run to see what needs to be changed, and vendor/bin/php-cs-fixer fix if you're OK with automatic changes.
@osiset please check the latest commits
Tests now pass, linting is good. @nahid
@Kyon147 I'm good with this one too if you are, seems like a feature thats been asked about before a few times.
Yeah looks good to me too
@nahid Please review once more, and if you're good we'll merge this in.
@nahid Please review once more, and if you're good we'll merge this in.
@osiset sure
| gharchive/pull-request | 2022-08-25T07:24:35 | 2025-04-01T06:45:19.085736 | {
"authors": [
"Kyon147",
"nahid",
"osiset"
],
"repo": "osiset/laravel-shopify",
"url": "https://github.com/osiset/laravel-shopify/pull/1192",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1162512314 | motd spam
Hi,
is there a reason why we ship a motd in roles/motd/defaults/main.yml?
This clutters the output we get from SSH:
WARNING *
You are accessing a secured system and your actions will be logged along *
with identifying information. Disconnect immediately if you are not an *
authorized user of this system. *
Filesystem Size Used Avail Use% Mounted on
or
WARNING *
You are accessing a secured system and your actions will be logged along *
with identifying information. Disconnect immediately if you are not an *
authorized user of this system. *
Linux hostname1 5.4.0-80-generic #90-Ubuntu SMP Fri Jul 9 22:49:44 UTC 2021 x86_64 x86_64 x86_64 GNU/Linux
IMHO anyone "unauthorized" to connect to the system would disconnect from it due to the above.
regards,
Ralf
This was/is the default in the ansible-hardening role. The note is there by default because some security certifications (STIG?) list it as requirement to have it. You can set the parameter to ““. This was you disable the message.
The operating system must display the Standard Mandatory DoD Notice and Consent Banner before granting local or remote access to the system via a graphical user logon. (V-71859)¶
Department of Defense probably does not apply to alot of us, but I'm fine if this is the reason why it's done.
We extracted the motd tasks from the ansible-hardening role and took the defaults from there. The ansible-hardening role implements all these requirements.
Let's discuss this briefly next week in SCS Team 1 how the others see it. I am open to set this to an empty string by default.
| gharchive/issue | 2022-03-08T10:57:22 | 2025-04-01T06:45:19.092701 | {
"authors": [
"berendt",
"frosty-geek"
],
"repo": "osism/ansible-collection-commons",
"url": "https://github.com/osism/ansible-collection-commons/issues/296",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2282780435 | [ceph-dashboard] Make standby behaviour and status code configurable
Allow to configure the ceph dashboard's options standby_behaviour ([1]) and standby_error_status_code ([2]).
While keeping the current default to redirect, this allows to adapt the behaviour to be more compatible with deployment behind a loadbalancer, where a redirect to a possibly internal URL is less useful than an error.
Especially in the case of a deployment behind haproxy one may set
ceph_dashboard_standby_behaviour: error
ceph_dashboard_standby_error_status_code: 404
and configure haproxy to
http-check expect status 200,404
http-check disable-on-404
where the former will set the backend "UP" on status codes 200 and 404, while latter will immediately transition the server to the "NOLB" status, thus not marking passive instances as down, while they are actually functioning correctly.
[1]
https://docs.ceph.com/en/quincy/mgr/dashboard/#disable-the-redirection
[2]
https://docs.ceph.com/en/quincy/mgr/dashboard/#configure-the-error-status-code
Part of https://github.com/osism/issues/issues/1013
| gharchive/pull-request | 2024-05-07T09:44:24 | 2025-04-01T06:45:19.096540 | {
"authors": [
"janhorstmann"
],
"repo": "osism/container-image-ceph-ansible",
"url": "https://github.com/osism/container-image-ceph-ansible/pull/481",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
98443059 | installation on Ubuntu
Running npm install osmium on Ubuntu 15.04 and node 0.10.25 gave the following error message:
...
> node-pre-gyp install --fallback-to-build
/usr/bin/env: node: No such file or directory
npm WARN This failure might be due to the use of legacy binary "node"
npm WARN For further explanations, please read
/usr/share/doc/nodejs/README.Debian
npm ERR! osmium@0.3.0 install: `node-pre-gyp install --fallback-to-build`
npm ERR! Exit status 127
npm ERR!
npm ERR! Failed at the osmium@0.3.0 install script.
npm ERR! This is most likely a problem with the osmium package,
npm ERR! not with npm itself.
npm ERR! Tell the author that this fails on your system:
npm ERR! node-pre-gyp install --fallback-to-build
npm ERR! You can get their info via:
npm ERR! npm owner ls osmium
npm ERR! There is likely additional logging output above.
npm ERR! System Linux 3.19.0-25-generic
npm ERR! command "/usr/bin/nodejs" "/usr/bin/npm" "install" "osmium"
npm ERR! cwd /home/azureuser
npm ERR! node -v v0.10.25
npm ERR! npm -v 1.4.21
npm ERR! code ELIFECYCLE
npm ERR!
npm ERR! Additional logging details can be found in:
npm ERR! /home/azureuser/npm-debug.log
npm ERR! not ok code 0
It seems that the Makefile uses the node-command, but apt-get install nodejs makes this available as nodejs. A workaround is to run sudo apt-get install nodejs-legacy, see this SO question. After this npm install osmium installed node-osmium (although with same error message as in issue #45).
apt-get install nodejs-legacy is the correct approach.
| gharchive/issue | 2015-07-31T18:06:02 | 2025-04-01T06:45:19.123853 | {
"authors": [
"matiasdahl",
"springmeyer"
],
"repo": "osmcode/node-osmium",
"url": "https://github.com/osmcode/node-osmium/issues/48",
"license": "BSL-1.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1994610635 | Using iavl v1.0.0 with cosmos sdk 0.47.5
Closes: #XXX
What is the purpose of the change
Add a description of the overall background and high-level changes that this PR introduces
(E.g.: This pull request improves documentation of area A by adding ....
Brief Changelog
(for example:)
The metadata is stored in the blob store on job creation time as a persistent artifact
Deployments RPC transmits only the blob storage reference
Daemons retrieve the RPC data from the blob cache
Testing and Verifying
(Please pick one of the following options)
This change is a trivial rework / code cleanup without any test coverage.
(or)
This change is already covered by existing tests, such as (please describe tests).
(or)
This change added tests and can be verified as follows:
(example:)
Added unit test that validates ...
Added integration tests for end-to-end deployment with ...
Extended integration test for ...
Manually verified the change by ...
Documentation and Release Note
Does this pull request introduce a new feature or user-facing behavior changes? (yes / no)
Is a relevant changelog entry added to the Unreleased section in CHANGELOG.md? (yes / no)
How is the feature or change documented? (not applicable / specification (x/<module>/spec/) / Osmosis docs repo / not documented)
Fyi, on the PR I made, we got a super helpful comment from @julienbrt
https://github.com/julienrbrt/cosmos-sdk-inj/compare/f/v0.47.3-inj-1...ac1b893f08942b97c308dc57716cf8c597ef967f
Closing, since we currently are using a different branch for this
| gharchive/pull-request | 2023-11-15T11:39:15 | 2025-04-01T06:45:19.138320 | {
"authors": [
"czarcas7ic",
"faddat",
"hieuvubk"
],
"repo": "osmosis-labs/cosmos-sdk",
"url": "https://github.com/osmosis-labs/cosmos-sdk/pull/495",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1909456985 | [VS Incentives]: Consider abstracting gauge logic into a general interface for extensibility
Background
Carried over from this discussion started by @p0mvn.
"I think one of the struggles with regular gauges is extending them with the new functionality. Since those were implemented as a concrete struct, we had to extend them by creating new fields resulting in a bloated abstraction.
To prevent that happening with GroupGauge in the future, do you think it makes sense to implement it as an interface (proto Any) and separate each implementation based on splitting policy.
This would make it easier to extend in the future if needed. It will also avoid having the code smell of deciding the execution flow based on splitting policy via if/else. Instead, we could implement a method on the group gauge. The concrete implementation would be chosen dynamically based on type."
Suggested Design
For example, this
func (k Keeper) syncGroupGaugeWeights(ctx sdk.Context, groupGauge types.GroupGauge) error {
if groupGauge.SplittingPolicy == types.Volume {
err := k.syncVolumeSplitGauge(ctx, groupGauge)
if err != nil {
return err
}
} else {
return types.UnsupportedSplittingPolicyError{GroupGaugeId: groupGauge.GroupGaugeId, SplittingPolicy: groupGauge.SplittingPolicy}
}
return nil
}
would become this:
volumeSplittingGauge := ...
volumeSplittingGauge.syncWeights(ctx)
evenSplittingGauge := ...
evenSplittingGauge.syncWeights(ctx)
Acceptance Criteria
Issue is discussed and converged on. If we decide to implement, then interface is implemented and well tested.
Thanks for tracking this @AlpinYukseloglu
On more thinking, while I think that this is a clean pattern overall, there are certain issues present that make it less preferred here.
For example, since syncWeights accesses store, it will also have to take in a keeper as parameter in addition to context.
Like this:
volumeSplittingGroup.syncWeights(ctx, k) // k is incentives keeper
Since Group is defined in types, any internal unexported method defined on the incentives keeper would be inaccessible and have to be exported.
As a result, I think we should be good to close this. Please let me know if disagreement
| gharchive/issue | 2023-09-22T20:02:42 | 2025-04-01T06:45:19.146790 | {
"authors": [
"AlpinYukseloglu",
"p0mvn"
],
"repo": "osmosis-labs/osmosis",
"url": "https://github.com/osmosis-labs/osmosis/issues/6499",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1098247610 | Osmosis Bash Installer V2
V2 should include (for a release in the next sprint)
[x] State Sync (I'm trying to hold out for the fix to be more permanent, but I can include the temp fix for now)
[x] Mac ARM compatibility
[x] Pruning methods (aggressive pruning, default, archive) (specifically make it so that users downloading archive don't prune)
[x] Custom ports
[ ] Replay from genesis (Will have to find a way to capture the chain halting due to version mismatch on python, update version, then restart)
[ ] Mac compatibility
[ ] Ability to choose installation directory (currently just installs default location)
[ ] Download releases
No longer needed (due to discussion with Dev):
[ ] Windows compatibility
Closing issue and moving to installer repo
| gharchive/issue | 2022-01-10T18:41:07 | 2025-04-01T06:45:19.150109 | {
"authors": [
"czarcas7ic"
],
"repo": "osmosis-labs/osmosis",
"url": "https://github.com/osmosis-labs/osmosis/issues/746",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
928250211 | Update Windows 4.9.0 hash after GH workflow was rerun
Heads up @directionless and @muffins, we need to coordinate re-running GH workflows after a release. When the S3 and tag assets are updated, the hash on the website needs to be updated too. It's best to only run those workflows once.
I think we'll slowly link all the workflows together
| gharchive/pull-request | 2021-06-23T13:18:42 | 2025-04-01T06:45:19.161622 | {
"authors": [
"directionless",
"theopolis"
],
"repo": "osquery/osquery-site",
"url": "https://github.com/osquery/osquery-site/pull/221",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
920017497 | OVC3 jumbo frames may cause instabilities
Documenting the issue so I don't forget to look into it.
OVC3 is currently configured to use jumbo frames in its ethernet interface:
ubuntu@arm:~$ cat /etc/dhcp/dhcpd.conf | grep -n 13500
45: option interface-mtu 13500;
This can cause some instabilities: the interface stops sending data in an unpredictable way. Commenting this line solves the issue.
Interesting, never noticed it before. I remember setting that number as high as possible since the lower it was the lower the performance of the USB ethernet was.
I'd suggest running a benchmark with iperf3 to see if it significantly affects performance and whether the tradeoff is acceptable for your application.
Hi @luca-della-vedova. Yes, I also noticed that this behavior depends significantly on the computer the OVC is connected to. I have observed that some computers are able to communicate and transmit simple messages, such as ssh commands (less than MTU?). But things get unstable when we start transmitting images and the connection just hangs.
As you suggested, it may be interesting to get some benchmarks on this issue.
Hi @fcladera
I was probing around the documentation of the Xilinx tools and found out that in the 2021.1 version of the tools released just a few weeks ago, I saw that in the changelog (under Linux kernel bugfixes) there is an entry:
Fixed USB3.0 RNDIS Gadget issue when MTU value is set to 15300.
I wonder if it could help, upgrading the tools to 2021.1 would bring that fix in but if that is too daunting it should be possible to just get the relevant patch and add it to the kernel compilation recipe in the petalinux project
| gharchive/issue | 2021-06-14T05:31:20 | 2025-04-01T06:45:19.166798 | {
"authors": [
"fcladera",
"luca-della-vedova"
],
"repo": "osrf/ovc",
"url": "https://github.com/osrf/ovc/issues/53",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2739256992 | Add 5 new terms
Continuing to build the new glossary, this PR adds terms for:
Attestation
Authentication
Authenticated attacker
Authorization
Availability
Added 10 more terms to this pull request
| gharchive/pull-request | 2024-12-13T21:45:30 | 2025-04-01T06:45:19.185223 | {
"authors": [
"GeauxJD"
],
"repo": "ossf/glossary",
"url": "https://github.com/ossf/glossary/pull/4",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
953261439 | Update links within docs
Links within the doc currently go to github.com/google/oss-vulnerability-guide instead of github.com/ossf.../
Resolved
| gharchive/issue | 2021-07-26T20:27:50 | 2025-04-01T06:45:19.186631 | {
"authors": [
"annabellegoth2boss"
],
"repo": "ossf/oss-vulnerability-guide",
"url": "https://github.com/ossf/oss-vulnerability-guide/issues/7",
"license": "CC-BY-4.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2631936638 | fix(redhat_osv): restructure test data
This commit restructures the test data so that CSAF and OSV data is not co-mingled in the same directory. It is an attept to address
$ go install github.com/ossf/osv-schema/tools/osv-linter@latest
go: downloading github.com/ossf/osv-schema v1.6.7
go: github.com/ossf/osv-schema/tools/osv-linter@latest: create zip: tools/redhat/testdata/rhsa-2024_4546.json: case-insensitive file name collision: "tools/redhat/testdata/RHSA-2024_4546.json" and "tools/redhat/testdata/rhsa-2024_4546.json"
@jasinner FYI, also what's the go with "test_date" here? https://github.com/ossf/osv-schema/blob/34c8f0fb9794582b655902acd47dc6ef7db6e0b8/tools/redhat/redhat_osv/osv_test.py#L30-L30
Looks good to me. If we wanted to match more closely with the filenames produced by the redhat_osv converter we could change the expected output filename to use a colon like 'RHSA-2024:6220.json'. I just didn't do that because the colon character needs to be escaped in a Linux shell.
test_date
what's the go with "test_date" here?
The short version is that the date is not used in the test. The long version is that the 'modified' date used in the OSV class is passed into the converter instead of being calculated based on the current time. We we do the conversion of the data we use the same modified date as the CSAF file because we that value to determine if the content has changed or not since the last update.
This has the unfortunate side affect that if the OSV data is regenerated with new conversion code the modified date of the OSV records will not be updated.
| gharchive/pull-request | 2024-11-04T06:14:29 | 2025-04-01T06:45:19.190140 | {
"authors": [
"andrewpollock",
"jasinner"
],
"repo": "ossf/osv-schema",
"url": "https://github.com/ossf/osv-schema/pull/308",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2196656835 | Add token price resolvers
Adds support for using UniswapV2 and UniswapV3 spot prices for price estimations, enabling tokens without Chainlink data feeds to have their price estimates shown.
In addition, this change allows the native token USD price data feed contract address to be specified in the config.
Note that the UniswapV3 "is there enough liquidity to trust this price" check is subject to being incorrect if there is a substantial amount of out-of-range liquidity and little or no liquidity in the active range, e.g. if the true price of the token is higher than all active liquidity positions. If it's easy to calculate how much liquidity is on either side of the current token price without having to make too many extra calls, that could be a future improvement.
Other future improvement: Users may wish to know the source of the price data. We may also wish to change the background color of the price estimate based on the source of the information.
Closes #1879.
I think you forgot to embed the configuration for chainId == 10 (op-mainnet), which was the main driver of this issue :)
in fact, I was not thinking about adding alternative resolvers on mainnet for now in order to not complicate things, but you rightfully guessed where we were aiming to :)
but yes, regarding L1 mainnet, one of the issues created by having more than 1 resolver is trust and transparency regarding the source. I was thinking about having difference colors for different sources, maybe we change the chainlink to use a "chainlink blue logo"-color and uni use a "uniswap-pink"-color?
Uniswap price sources have a fuschia color. I decided to keep the neutral gray color for Chainlink price sources since I didn't want the price information to become too much of a distraction.
For comparison, it would look like this with a blue color:
Optimism now has its price oracle information embedded without the need for a custom configuration.
Uniswap price sources have a fuschia color. I decided to keep the neutral gray color for Chainlink price sources since I didn't want the price information to become too much of a distraction.
ok, let's keep the gray considering Chainlink is the "canonical" oracle. but open to revisit it.
| gharchive/pull-request | 2024-03-20T04:46:29 | 2025-04-01T06:45:19.232055 | {
"authors": [
"sealer3",
"wmitsuda"
],
"repo": "otterscan/otterscan",
"url": "https://github.com/otterscan/otterscan/pull/1899",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
939046599 | homework 1
Выполнено ДЗ № 1
[x] Основное ДЗ
[x] Задание со *
В процессе сделано:
Подготовлен Dockerfile, готовый образ запушен в публичный Container Registry (Docker Hub)
Создан манифест первого пода, добавлен init-контейнер для генерации index.html
Запуск микросервиса frontend, исправление ошибки при старте пода
Как запустить проект:
В директории kubernetes-intro выполнить:
kubectl apply -f web-pod.yaml
kubectl port-forward --address 0.0.0.0 pod/web-app 8000:8000
Как проверить работоспособность:
Перейти по ссылке http://localhost:8000
PR checklist:
[x] Выставлен label с темой домашнего задания
Добрый день! Задание выполнено верно
| gharchive/pull-request | 2021-07-07T16:07:45 | 2025-04-01T06:45:19.243372 | {
"authors": [
"DmitryMCN",
"Konstantinov86"
],
"repo": "otus-kuber-2021-06/DmitryMCN_platform",
"url": "https://github.com/otus-kuber-2021-06/DmitryMCN_platform/pull/6",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1292375569 | kubernetes-security
Выполнено ДЗ №4
[v] Основное ДЗ
[отсутствуют] Задание со *
В процессе сделано:
Выполнены все задания ДЗ
Как запустить проект:
Применить манифесты, упорядоченный по директориям в соответствии с условиями ДЗ
Как проверить работоспособность:
Один из вариантов проверки - выполнить какое-либо действие от имени сервисного аккаунта, созданного в рамках задания ДЗ.
Например, для проверки 5 пункта 3-го задания (по условиям, у сервисного аккаунта ken в неймспейсе dev есть только права на чтение ресурсов в этом неймспейсе) можно выполнить следующую команду:
проверка наличия прав на создание ресурса pod:
kubectl auth can-i create pods -n dev --as=system:serviceaccount:dev:ken
проверка прав на чтение ресурса pod в неймспесе dev:
kubectl auth can-i get pods -n dev --as=system:serviceaccount:dev:ken
проверка прав на чтение ресурса pod в неймспесе default:
kubectl auth can-i get pods --as=system:serviceaccount:dev:ken
PR checklist:
[v] Выставлен label с темой домашнего задания
Добрый день ! Дз выполнено верно
| gharchive/pull-request | 2022-07-03T17:27:36 | 2025-04-01T06:45:19.247868 | {
"authors": [
"BorshchevY",
"Konstantinov86"
],
"repo": "otus-kuber-2022-06/BorshchevY_platform",
"url": "https://github.com/otus-kuber-2022-06/BorshchevY_platform/pull/4",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
930799376 | Adding the i18n keys for the voucher status
Screenshots
English
French
Needed for the pull request #175 of the frontend project :)
I've removed the extra space, thank you!
Let's make sure to get a review before merging in the future :)
@ann-kilzer My mistake, I'll do that next time.
| gharchive/pull-request | 2021-06-26T21:46:23 | 2025-04-01T06:45:19.255742 | {
"authors": [
"JohnTitor2036",
"ann-kilzer"
],
"repo": "ourjapanlife/findadoc-localization",
"url": "https://github.com/ourjapanlife/findadoc-localization/pull/68",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1270582701 | Bug in Laravel Octane
Hi! In this point http://joxi.ru/8An8ERqcNkl3Om we have the same data between requests and memory leaks when we use Laravel Octane.
Hmm, I'm not even sure how to fix this to be honest. Maybe it's possible to unset the listener somehow?
Confirmed!
In Octane you should not use such model static methods because of memory leaks. So multiselect field will execute this part of code (model::saved event) on each model change (e.g. reorder, or something else which is not related to multiselect relation).
As a simple workaround (without package modification) is to add custom fillUsing callback:
Multiselect::make(__('teams/team-user.fields.role'), 'roles')
->nullable()
->belongsToMany(TeamRoleResource::class, false) // ! order is important
->fillUsing(fn (NovaRequest $request, TeamUser $model, string $attribute, string $requestAttribute) => $model->{$attribute}()?->sync($request->get($requestAttribute) ?: []))
Heya! Thanks for the report. Should now be fixed in version 4.3.5. ;)
Cool, thanks!
Waiting for release!
| gharchive/issue | 2022-06-14T10:19:18 | 2025-04-01T06:45:19.258789 | {
"authors": [
"LikoSoftDev",
"Tarpsvo",
"pavloniym"
],
"repo": "outl1ne/nova-multiselect-field",
"url": "https://github.com/outl1ne/nova-multiselect-field/issues/160",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
355852031 | 神超大神,我是第7000个stars,非常喜欢您。
like you very much!
what happened ?
| gharchive/issue | 2018-08-31T06:12:12 | 2025-04-01T06:45:19.314228 | {
"authors": [
"danic-io",
"zengshiran"
],
"repo": "overtrue/wechat",
"url": "https://github.com/overtrue/wechat/issues/1352",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1221223242 | Update wiremock-jre8 to 2.33.2
Updates com.github.tomakehurst:wiremock-jre8 from 2.32.0 to 2.33.2.
GitHub Release Notes - Version Diff
I'll automatically update this PR to resolve conflicts as long as you don't change it yourself.
If you'd like to skip this version, you can just close this PR. If you have any feedback, just mention me in the comments below.
Configure Scala Steward for your repository with a .scala-steward.conf file.
Have a fantastic day writing Scala!
Adjust future updates
Add this to your .scala-steward.conf file to ignore future updates of this dependency:
updates.ignore = [ { groupId = "com.github.tomakehurst", artifactId = "wiremock-jre8" } ]
Or, add this to slow down future updates of this dependency:
dependencyOverrides = [{
pullRequest = { frequency = "@monthly" },
dependency = { groupId = "com.github.tomakehurst", artifactId = "wiremock-jre8" }
}]
labels: library-update, early-semver-minor, semver-spec-minor, commit-count:1
Superseded by #202.
| gharchive/pull-request | 2022-04-29T16:36:37 | 2025-04-01T06:45:19.348692 | {
"authors": [
"scala-steward"
],
"repo": "ovotech/kafka-serialization",
"url": "https://github.com/ovotech/kafka-serialization/pull/199",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
956328579 | escrow authorization: authorize deployment payment
Overview
We would like to be able for an account to be able to pay for deployments owned by any number of other accounts.
The first use-case for this will be to give users free hosting for some amount of time, without transferring them tokens.
Background
Authz
Newer versions of cosmos-sdk include an authz mechanism that allows for one account to authorize actions for other accounts.
Escrow
Akash uses a so-called escrow accounts for lease payments. This escrow mechanism is implemented in a standalone module x/escrow.
The escrow account is currently funded from the account that owns the deployment. Funds are transferred from the owner account to the escrow account in two transactions:
when creating the deployment
when "topping up" the escrow account.
When a deployment is closed, any balance remaining in the escrow account is currently returned to the owning account.
Requirements
Create a set of authorization objects that can be used to allow a separate account to both fund the deployment escrow account and receive remaining balances from the escrow account when the deployment is closed.
funding account must be able to fund deployments for any number of other accounts.
funding account must be able to declare maximum limit of funding for each account.
funding account must be able to de-authorize any unused funding amount.
an escrow account may have only one non-owner funding account
deposits to the escrow account after deployment creation can be made by either the owning account or the authorized funding account.
Note, to begin with, this authorization is only for authorizing spending - the deployment owner controls all other aspects of the deployment, including:
the funding account cannot close the deployment.
the funding account cannot restrict which providers are able win leases for the deployment.
CLI
Create a set of akash tx, akash query commands that enable, use, and report on deployment payment authorizations.
Example workflow:
# allow "deploy-wallet" to use 20AKT for deployments from "funding-wallet"
akash tx escrow authorize $(akash keys show deploy-wallet -a) 20000000uakt --from funding-wallet
# create a deployment from "deploy-wallet" that is paid for by "funding-wallet"
akash tx deployment create --payment-from $(akash keys show funding-wallet -a) --from deploy-wallet
# view authorizations created by the funding wallet
akash query escrow authorizations $(akash keys show funding-wallet -a)
# reduce amount for an authorization
akash tx escrow update $(akash keys show deploy-wallet -a) 10000000uakt --from funding-wallet
# deposit additional funds from the owning account
akash tx deployment deposit --dseq 1 --from deployment-wallet
Super excited about this!
Question: It might happen that an escrow account can get funded by both the funder and the deployment owner. Let's say while creating the escrow account:
it was funded by the funder with 5akt
then 3akt were deducted from it as provider's fees
then the deployment owner added 10akt to the escrow account
2akt got used again as provider's fees
then the deployment owner decided to cancel the lease, so the escrow account needs to be closed.
In this situation, the escrow account has 10akt left. The funding wallet had only sent 5akt to the escrow while the deployment owner had sent 10akt. If we send the remaining 10akt to the funding wallet, then it doesn't seem fair. What should we do?
Possible Solution:
In escrow account for app deployment we maintain two seperate balance.
Direct fund(Pays from deployment wallet)
Allowance(Pays from funding wallet)
We prioritize spending from Allowance and then use Direct fund.
In case of lease cancel, we return back Direct fund to the Deployment wallet and Allowance to Funding wallet.
Maintaining two separate balances could work, but only if there was a single external funding account. I think we want any number of external accounts to fund any single deployment. Is that accurate @boz ?
It's okay if we have just one external funding account. if it can be made more general and allow for multiple in a clean way* then that would be great.
*maybe always have one active payer, and order by (external?,time created).
In that case the solution proposed by @arijitAD sounds good to me.
| gharchive/issue | 2021-07-30T01:48:47 | 2025-04-01T06:45:19.361016 | {
"authors": [
"arijitAD",
"boz",
"gosuri",
"hydrogen18"
],
"repo": "ovrclk/akash",
"url": "https://github.com/ovrclk/akash/issues/1344",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2052858229 | Declare variables required for report
Sometimes there are
Traceback (most recent call last):
File "/usr/local/bin/depscan", line 8, in <module>
sys.exit(main())
File "/usr/local/lib/python3.9/site-packages/depscan/cli.py", line 1118, in main
vdr_file=vdr_file,
UnboundLocalError: local variable 'vdr_file' referenced before assignment
errors on depscan executions, thus declaring all variables used for the
reporting that are initialized only conditionally upfront.
Was pylint not capturing this?
how to run this locally, or find issues reported by GH actions?
@heubeck
pip install pylint
pylint cli.py
If you use pycharm, you'll also get a lot of stuff like this in the Problems tab, and even more if you install the sourcery plugin (which is free to use with open source projects). The qodana local report feature is also nice, but not sure if that's included in the free version of pycharm.
| gharchive/pull-request | 2023-12-21T18:02:20 | 2025-04-01T06:45:19.367912 | {
"authors": [
"cerrussell",
"heubeck"
],
"repo": "owasp-dep-scan/dep-scan",
"url": "https://github.com/owasp-dep-scan/dep-scan/pull/191",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2253305159 | Phasing out SecStatusEngine
A small discussion happened on the Slack #project-modsecurity from OWASP where I pointed out that, with TW changing ownership to OWASP of modsecurity, the domain name might need to be transmitted so that the SecStatusEngine option (https://github.com/owasp-modsecurity/ModSecurity/wiki/Reference-Manual-(v2.x)#secstatusengine), if this is still used, is still working as expected.
According to @fzipi, this option should be disabled since a long time ago. This has been done in the default config 2 months ago by @airween (https://github.com/owasp-modsecurity/ModSecurity/commit/f850932f83d47a68137ab207e6db0f217152c707).
@dune73 mentioned that this domain will be still in the hands of TW until Summer 2024, and that TW is not having their status engine in operation for quite some time.
That being said, we all pretty much agree that:
TW status engine will not come back.
This is now disabled by default, so very little people are going to turn it on.
This option will not be useful anymore, and will just pollute (if enabled) the DNS.
This options is not supported in v3 (https://github.com/owasp-modsecurity/ModSecurity/wiki/Reference-Manual-(v3.x)#secstatusengine), and will not be created.
Functionally, it does not bring anything to the end user.
This is where I'm proposing removal of this option from v2, while knowing this operation should be carefully considered so that no configuration gets broken.
Probably we could first warn about this option being deprecated, following by removing the actual logic while keeping the warning, and finally removing this option altogether from the parsing logic and the documentations.
Hi @vloup, thanks for this suggestion. I think this is a useful plan, and a good idea.
I think we can manage to do two steps in one, and that sounds better to me as well:
Deprecate this option and add a warning message.
We directly remove all the code related to this option.
Added https://github.com/owasp-modsecurity/ModSecurity/pull/3157
@marcstern do you think we can close this issue?
| gharchive/issue | 2024-04-19T15:36:13 | 2025-04-01T06:45:19.374089 | {
"authors": [
"airween",
"marcstern",
"vloup"
],
"repo": "owasp-modsecurity/ModSecurity",
"url": "https://github.com/owasp-modsecurity/ModSecurity/issues/3122",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
124237301 | It needs to be possible to un-fav from fav section
This is a comment mentioned by Colin during the meeting.
Meeting today? Or a previous one. I was thinking we should have a system for this, but it perhaps needs to wait on a UX design?
UX design would be needed, for now we can move it to a later release?
This includes:
There's currently a branch with the notification symbol, and a favourite star implemented on the search tab. I'd be in favour of not duplicating this code so if we can create some sort of toggle-icon that can be re-used that would be good.
Closed by #137. Haven't implemented the designed confirmation window yet, but using the built-in confirm window.
| gharchive/issue | 2015-12-29T16:50:14 | 2025-04-01T06:45:19.376818 | {
"authors": [
"bjedrzejewski",
"owennw"
],
"repo": "owennw/OpenFinD3FC",
"url": "https://github.com/owennw/OpenFinD3FC/issues/120",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
193099989 | Cannot use on Allegro mlisp
See https://github.com/edicl/cl-unicode/issues/10
cl-unicode works with alisp, but agraph-client requires mlisp.
It worked now with
$ mlisp
International Allegro CL Professional Edition
10.0 [Mac OS X (Intel)] (Apr 30, 2018 18:57)
Copyright (C) 1985-2015, Franz Inc., Oakland, CA, USA. All Rights Reserved.
This development copy of Allegro CL is licensed to:
[TC21830] IBM Research Lab - Brazil
...
cl-user(1): (ql:quickload :cl-unicode)
To load "cl-unicode":
Load 1 ASDF system:
cl-unicode
; Loading "cl-unicode"
(:cl-unicode)
@fcbr do you agree on close this issue?
| gharchive/issue | 2016-12-02T11:58:07 | 2025-04-01T06:45:19.401852 | {
"authors": [
"arademaker",
"fcbr"
],
"repo": "own-pt/cl-conllu",
"url": "https://github.com/own-pt/cl-conllu/issues/4",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
2533295511 | natsjskv store needs a migration from 5 to 6
when using nats-js-kv, eg using the helm charts you might run into this error:
2024-09-17T14:09:40+02:00 ERR error uploading file error="Failed to store data in bucket 'spaces/f1/bdd61a-da7c-49fc-8203-0558109d1b4f/nodes/64/bb/19/df/-1598-4691-a07a-d2c2f5714815.REV.2024-09-17T08:56:50.792574043Z': nats: invalid key" datatx=simple line=/home/jfd/Repositories/reva/pkg/rhttp/datatx/manager/simple/simple.go:147 pkg=rhttp service=storage-system traceid=49388ac282e8ffca3a5165cb484669b1
this is caused by the store now encoding keys.
we need a migration for this because otherwise not only cached file metadata becomes inaccessible (this will just phase out) but because other stores use it to persist data (not as a cache).
stable 5 uses the OCIS_CACHE_STORE env var for these services:
./services/frontend/pkg/config/config.go: StatCacheType string `yaml:"stat_cache_type" env:"OCIS_CACHE_STORE;FRONTEND_OCS_STAT_CACHE_STORE" desc:"The type of the cache store. Supported values are: 'memory', 'redis-sentinel', 'nats-js-kv', 'noop'. See the text description for details." introductionVersion:"pre5.0"`
./services/gateway/pkg/config/config.go: ProviderCacheStore string `yaml:"provider_cache_store" env:"OCIS_CACHE_STORE;GATEWAY_PROVIDER_CACHE_STORE" desc:"The type of the cache store. Supported values are: 'memory', 'redis-sentinel', 'nats-js-kv', 'noop'. See the text description for details." introductionVersion:"pre5.0"`
./services/gateway/pkg/config/config.go: CreateHomeCacheStore string `yaml:"create_home_cache_store" env:"OCIS_CACHE_STORE;GATEWAY_CREATE_HOME_CACHE_STORE" desc:"The type of the cache store. Supported values are: 'memory', 'redis-sentinel', 'nats-js-kv', 'noop'. See the text description for details." introductionVersion:"pre5.0"`
./services/settings/pkg/config/config.go: Store string `yaml:"store" env:"OCIS_CACHE_STORE;SETTINGS_CACHE_STORE" desc:"The type of the cache store. Supported values are: 'memory', 'redis-sentinel', 'nats-js-kv', 'noop'. See the text description for details." introductionVersion:"pre5.0"`
./services/graph/pkg/config/cache.go: Store string `yaml:"store" env:"OCIS_CACHE_STORE;GRAPH_CACHE_STORE" desc:"The type of the cache store. Supported values are: 'memory', 'redis-sentinel', 'nats-js-kv', 'noop'. See the text description for details." introductionVersion:"pre5.0"`
./services/storage-users/pkg/config/config.go: Store string `yaml:"store" env:"OCIS_CACHE_STORE;STORAGE_USERS_FILEMETADATA_CACHE_STORE" desc:"The type of the cache store. Supported values are: 'memory', 'redis-sentinel', 'nats-js-kv', 'noop'. See the text description for details." introductionVersion:"pre5.0"`
./services/storage-users/pkg/config/config.go: Store string `yaml:"store" env:"OCIS_CACHE_STORE;STORAGE_USERS_ID_CACHE_STORE" desc:"The type of the cache store. Supported values are: 'memory', 'redis-sentinel', 'nats-js-kv', 'noop'. See the text description for details." introductionVersion:"pre5.0"`
./services/ocs/pkg/config/config.go: Store string `yaml:"store" env:"OCIS_CACHE_STORE;OCS_PRESIGNEDURL_SIGNING_KEYS_STORE" desc:"The type of the signing key store. Supported values are: 'redis-sentinel' and 'nats-js-kv'. See the text description for details." introductionVersion:"5.0"`
./services/proxy/pkg/config/config.go: Store string `yaml:"store" env:"OCIS_CACHE_STORE;PROXY_OIDC_USERINFO_CACHE_STORE" desc:"The type of the cache store. Supported values are: 'memory', 'redis-sentinel', 'nats-js-kv', 'noop'. See the text description for details." introductionVersion:"pre5.0"`
./services/proxy/pkg/config/config.go: Store string `yaml:"store" env:"OCIS_CACHE_STORE;PROXY_PRESIGNEDURL_SIGNING_KEYS_STORE" desc:"The type of the signing key store. Supported values are: 'redis-sentinel', 'nats-js-kv' and 'ocisstoreservice' (deprecated). See the text description for details." introductionVersion:"5.0"`
./services/storage-system/pkg/config/config.go: Store string `yaml:"store" env:"OCIS_CACHE_STORE;STORAGE_SYSTEM_CACHE_STORE" desc:"The type of the cache store. Supported values are: 'memory', 'redis-sentinel', 'nats-js-kv', 'noop'. See the text description for details." introductionVersion:"pre5.0"`
The OCIS_PERSISTENT_STORE env var is used for
./services/postprocessing/pkg/config/config.go: Store string `yaml:"store" env:"OCIS_PERSISTENT_STORE;POSTPROCESSING_STORE" desc:"The type of the store. Supported values are: 'memory', 'ocmem', 'etcd', 'redis', 'redis-sentinel', 'nats-js', 'noop'. See the text description for details." introductionVersion:"pre5.0"`
./services/eventhistory/pkg/config/config.go: Store string `yaml:"store" env:"OCIS_PERSISTENT_STORE;EVENTHISTORY_STORE" desc:"The type of the store. Supported values are: 'memory', 'ocmem', 'etcd', 'redis', 'redis-sentinel', 'nats-js', 'noop'. See the text description for details." introductionVersion:"pre5.0"`
./services/userlog/pkg/config/config.go: Store string `yaml:"store" env:"OCIS_PERSISTENT_STORE;USERLOG_STORE" desc:"The type of the store. Supported values are: 'memory', 'ocmem', 'etcd', 'redis', 'redis-sentinel', 'nats-js', 'noop'. See the text description for details." introductionVersion:"pre5.0"`
this came already with ocis v3.0.0:
Change - Updated Cache Configuration: #5829
We updated all cache related environment vars to more closely follow the go
micro naming pattern: - {service}_CACHE_STORE_TYPE becomes
{service}_CACHE_STORE or {service}_PERSISTENT_STORE -
{service}_CACHE_STORE_ADDRESS(ES) becomes {service}_CACHE_STORE_NODES - The
mem store implementation name changes to memory - In yaml files the cache
type becomes store We introduced redis-sentinel as a store implementation.
https://github.com/owncloud/ocis/pull/5829
The natsjskv encoding issue was fixed released with v6.0.0.
Before v6 nats-js-kv could not be used as a storage file metadata cache. It defaults to memory and had to be disabled in HA deployments using STORAGE_USERS_FILEMETADATA_CACHE_STORE=noop and STORAGE_SYSTEM_CACHE_STORE=noop.
Upgading does NOT need a migration. In fact, the helm chart can now use nats-js-kv instead of noop.
Also see https://github.com/owncloud/ocis-charts/issues/214
Ahhh, IIRC we started to also cache trash or revision metadata ... that introduced : in the key. Let me see if i can find the commit,,,
comparing the two lines
https://github.com/owncloud/ocis/issues/9114
2024-05-08T17:12:59+02:00 ERR error reading blobsize xattr, using 0 error="error reading blobsize xattr: Failed to store data in bucket 'spaces/so/me-admin-user-id-0000-000000000000/nodes/04/a5/b4/86/-4ce3-492e-a98e-87eddada0146.REV.2024-05-08T14:54:24.01526906Z': nats: invalid key" name=-4ce3-492e-a98e-87eddada0146.REV.2024-05-08T14:54:24.01526906Z pkg=rgrpc service=storage-users traceid=a8a634abcc56ef9bb072699c85e5de3e
this issue
2024-09-17T14:09:40+02:00 ERR error uploading file error="Failed to store data in bucket 'spaces/f1/bdd61a-da7c-49fc-8203-0558109d1b4f/nodes/64/bb/19/df/-1598-4691-a07a-d2c2f5714815.REV.2024-09-17T08:56:50.792574043Z': nats: invalid key" datatx=simple line=/home/jfd/Repositories/reva/pkg/rhttp/datatx/manager/simple/simple.go:147 pkg=rhttp service=storage-system traceid=49388ac282e8ffca3a5165cb484669b1
solved with https://github.com/cs3org/reva/pull/4862
no magration needed
| gharchive/issue | 2024-09-18T10:04:57 | 2025-04-01T06:45:19.638691 | {
"authors": [
"butonic"
],
"repo": "owncloud/ocis",
"url": "https://github.com/owncloud/ocis/issues/10098",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1805029815 | traces show wrong service names
The service names in tracing are wrong after merging some of the trace PRs:
The IdP has nothing to do with ListStorageSpaces, not as a client, not as a server ...
this only seems to happen when running in single process ...
@ainmosni @dragonchaser Can you fix that?
I'm not able to reproduce this in the single process mode.
closing, could not reproduce
| gharchive/issue | 2023-07-14T15:00:45 | 2025-04-01T06:45:19.641861 | {
"authors": [
"ainmosni",
"butonic",
"dragonchaser",
"micbar"
],
"repo": "owncloud/ocis",
"url": "https://github.com/owncloud/ocis/issues/6813",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2296822989 | [stable] nightly failure on stable 5.0 on e2e tests
And "Alice" navigates to the projects space page
And "Alice" navigates to the project space "team.1"
And "Alice" uploads the following resources via drag-n-drop
│ resource │
│ simple.pdf │
│ testavatar.jpg │
✖ failed
locator.click: Timeout 30000ms exceeded.
=========================== logs ===========================
waiting for locator('#close-upload-info-btn')
============================================================
at Object.dropUploadFiles (/drone/src/webTestRunner/tests/e2e/support/objects/app-files/resource/actions.ts:580:45)
at async Resource.dropUpload (/drone/src/webTestRunner/tests/e2e/support/objects/app-files/resource/index.ts:39:5)
at async World.<anonymous> (/drone/src/webTestRunner/tests/e2e/cucumber/steps/ui/resources.ts:716:5)
And "Alice" creates a public link for the space with password "%public%" using the sidebar panel
The nightly fails on uploading via drag-n-drop step
Failure build: https://drone.owncloud.com/owncloud/ocis/34983/66/9
Fixed with PRs https://github.com/owncloud/web/pull/10910 and https://github.com/owncloud/ocis/pull/9175. So, closing this issue.
| gharchive/issue | 2024-05-15T04:11:02 | 2025-04-01T06:45:19.644384 | {
"authors": [
"PrajwolAmatya",
"S-Panta"
],
"repo": "owncloud/ocis",
"url": "https://github.com/owncloud/ocis/issues/9170",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2143147675 | refactor: Add more types and remove unnecessary .then()
Hi,
I added a few more type definitions (especially to data returned from axios) and a few other improvements.
Feel free to modify this PR if you don't like some of the changes.
Thanks for the plugin and keep up the good work!
Hey thanks! @Dunqing would you prefer to write a few test for generate-rules before validating these changes?
@Dunqing could you give a reason why this was closed?
@Dunqing could you give a reason why this was closed?
I apologize, we have completely refactored the current implementation in https://github.com/oxc-project/eslint-plugin-oxlint/pull/64. Your PR is outdated, so I closed it. Please feel free to send an PR to improve the current implementation. Thanks!
| gharchive/pull-request | 2024-02-19T21:11:41 | 2025-04-01T06:45:19.708205 | {
"authors": [
"Djiit",
"Dunqing",
"ffflorian"
],
"repo": "oxc-project/eslint-plugin-oxlint",
"url": "https://github.com/oxc-project/eslint-plugin-oxlint/pull/15",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2474982491 | feat(codegen)!: remove const generic MINIFY
This is a premature optimization, makes the code complicated, and bloats the final binary size.
The minify option is moved to CodegenOptions
CodSpeed Performance Report
Merging #5001 will not alter performance
Comparing codegen-remove-const-generic (da469d7) with main (b2ff2df)
Summary
✅ 29 untouched benchmarks
As I anticipated, negligible performance regression.
| gharchive/pull-request | 2024-08-20T08:02:25 | 2025-04-01T06:45:19.711182 | {
"authors": [
"Boshen"
],
"repo": "oxc-project/oxc",
"url": "https://github.com/oxc-project/oxc/pull/5001",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1879460457 | Fix 404 zlib dep source
Dependency has moved and was re-packaged. Old source returned 404, thus the build did not compile. Same story as https://github.com/oxen-io/oxen-core/pull/1653.
Fixed by 53d10eddeecae
| gharchive/pull-request | 2023-09-04T03:55:04 | 2025-04-01T06:45:19.712467 | {
"authors": [
"jagerman",
"martin-braun"
],
"repo": "oxen-io/oxen-storage-server",
"url": "https://github.com/oxen-io/oxen-storage-server/pull/481",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1093193123 | Implementing Lists in Rust
Hi there, I'd like to contribute by adding the DSAs in Rust, but I want to start with the Lists first as it is the simplest, I think. If not, you could direct me to it.
Thank you.
Hey, that sounds good
| gharchive/issue | 2022-01-04T10:04:30 | 2025-04-01T06:45:19.750459 | {
"authors": [
"Lilit0x",
"oyekanmiayo"
],
"repo": "oyekanmiayo/data-structures-all-langs",
"url": "https://github.com/oyekanmiayo/data-structures-all-langs/issues/48",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1846026769 | UX improvements
+ Changed from automatically searching .json files in BugFinding, let user select one or multiple json files
+ Expanded collapsible nodes by default on launching visualizer. Sometimes the Bug is collapsed because it is tied to a sequence of individual logs.
+ Larger dialog box to show more information
Nice, this is great work!
| gharchive/pull-request | 2023-08-10T23:48:25 | 2025-04-01T06:45:19.794551 | {
"authors": [
"ankushdesai",
"haoran-wen"
],
"repo": "p-org/peasy-ide-vscode",
"url": "https://github.com/p-org/peasy-ide-vscode/pull/30",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
127880962 | Add random baggel command along with parameters
This PR adds an endpoint to the Slack command, to allow people to order random flavoured bagels.
+1
| gharchive/pull-request | 2016-01-21T09:29:21 | 2025-04-01T06:45:19.830028 | {
"authors": [
"pCyril",
"quentinfayet"
],
"repo": "pCyril/devexcuses",
"url": "https://github.com/pCyril/devexcuses/pull/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1266265876 | A couple CLI tweaks
Hi @lauraluebbert ,
I noticed with gget 0.2.0 that gget --help returns with an error code. Also, it appears that gget -v or gget --version aren't currently working as expected.
Best,
Mike
Hi Michael,
The cli bugs should be fixed in version 0.2.1 (pip install --upgrade gget). Thank you very much for bringing this to my attention!
Best,
Laura
| gharchive/issue | 2022-06-09T14:58:33 | 2025-04-01T06:45:19.879655 | {
"authors": [
"lauraluebbert",
"mjsteinbaugh"
],
"repo": "pachterlab/gget",
"url": "https://github.com/pachterlab/gget/issues/16",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1938210470 | Feature proposal: KEGG PATHWAY image generation
Hello.
Thank you very much for developing this useful software.
I would like to propose a feature in gget enrichr that produces a KEGG PATHWAY image with the genes from the enrichment analysis highlighted.
Example usage:
gget enrichr -db pathway --kegg out.png --kegg_rank 1 DDX41 IRF3
will output an image (out.png) with IRF3 highlighted in the rank 1 pathway.
The feature uses the library pykegg. It checks if the package is installed and raises an error if not.
I would be grateful if you could review the PR.
@lauraluebbert Thank you very much for merging the PR!
| gharchive/pull-request | 2023-10-11T16:27:12 | 2025-04-01T06:45:19.881798 | {
"authors": [
"noriakis"
],
"repo": "pachterlab/gget",
"url": "https://github.com/pachterlab/gget/pull/106",
"license": "BSD-2-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
955318160 | Unclear How to Use Amazon RDS for the Postgres Server
Because of issue #6634 I tried to set up Postgres on Amazon RDS and use that instead of the internal Postgres instance. Looking at the logs, I could see that pg-bouncer was connecting to my RDS instance successfully, but then it was failing because the "pachyderm" database did not exist. I was expecting that it would automatically create the database.
After several failed attempts to connect to the database, pg-bouncer then started trying to use the user "pachyderm" to connect, rather than the user I configured. This failed because user "pachyderm" doesn't exist on my Postgres instance. It repeatedly tried to log in every second using this invalid user until I ran helm uninstall.
YAML file and logs attached:
helm-deploy.txt
postgres-errors.log
I changed the database name in my Helm chart from "pachyderm" to "postgres" ("postgres" being the name of my user account) and that got past the issue of the database not being found, EXCEPT that then it complained about not being able to find the "dex" database (which seems to be something not specified in the Helm YAML?)
So I created the "dex" database, and now everything starts up correctly.
I think there are two things here that need to be addressed:
Should we expect pg-bouncer to be able to create the "pachyderm" and "dex" databases if they don't exist? If yes, this is a bug that should be addressed (probably post-2.0). If no, then this is a docs issue; we need to clearly document the requirements for customers who wish to use their own Postgres instance.
If pg-bouncer is unable to connect, then: a) it probably shouldn't keep trying every second forever; b) it definitely shouldn't start trying to use the "pachyderm" user account rather than the user account that was configured. Again, I think this could be fixed post-2.0
For 1, I think this is probably a docs issue. In the case of self hosted postgres, the user should be responsible for setting up the user role and databases.
For 2, this is likely a bug, we likely missed a hardcoded instances of "pachyderm" in the db code.
Related: https://github.com/pachyderm/pachyderm/issues/6510
@chainlink - I agree on number 1; docs issue was the direction I was leaning as well. For number 2, it's weird because it used the correct username for several tries before it started trying to use the "pachyderm" username. Almost like someone thought, well if the configured name doesn't work let's try the default. Which isn't necessarily a bad fallback approach, except for the infinite retries. :)
It's possible that one connection path was through pg_bouncer (which passes through the user as configured), and one was direct. One could also have been the worker, which is improperly configured. We don't change the database configuration during any of the retry loops.
There are no places in the code where we call dbutil.NewDB without setting the user and database. We only default when the environment variable is not set. So it has to be an issue passing the environment variable to the worker, or getting it from the helm chart. It could also be an issue setting up dex, which uses different environment variables.
I think this could use a retest actually (I believe this was before we moved everything over to env vars)
We use an external PG instance on Hub, so it should work on AWS as well
The external PG instance works on AWS as long as:
The "dex" database has been created; and
You a.) use the default "postgres" login and "postgres" database for Pachyderm OR b.) create the specific user account and/or database you want to be used.
It's reasonable to expect the customer to set this up; we just need to make sure it gets into the docs. <-- @nadegepepin :)
Now that this is documented - Can we close this issue? https://docs.pachyderm.com/latest/deploy-manage/deploy/aws-deploy-pachyderm/
Or is there something pending still?
| gharchive/issue | 2021-07-28T23:57:35 | 2025-04-01T06:45:19.895088 | {
"authors": [
"brendoncarroll",
"chainlink",
"dgeorg42",
"nadegepepin"
],
"repo": "pachyderm/pachyderm",
"url": "https://github.com/pachyderm/pachyderm/issues/6639",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
231157432 | update word count to new format and corresponding docs
The word count example was still using the old format inputs. I updated the pipeline specs. I also did the following:
Removed the log, run2.sh, and run.sh files. Not sure why these were here. I know that the *.sh must not be in use, because they aren't even valid (they don't reference the right files or pipeline names).
Add a DAG figure to help people understand the flow.
Updated the README to improve follow and correct mistakes in a couple places.
Standardized and shortened the pipeline file names.
LGTM.
| gharchive/pull-request | 2017-05-24T20:12:15 | 2025-04-01T06:45:19.897827 | {
"authors": [
"derekchiang",
"dwhitena"
],
"repo": "pachyderm/pachyderm",
"url": "https://github.com/pachyderm/pachyderm/pull/1874",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
295948614 | Aren't there two sets of p_a parameters?
Bottom of p39 says there are p_a for the fishery and p_a for the survey.
So does that mean there are actually two p_a^{high} and p_a^{low} values? There may not be.
And then shouldn't (8) be summed over the fishery and the survey? Or, to save adding extra notation, this can just be explicitly mentioned.
Also, okay if I change p_a^{high} to just p_{\mbox{high}}. I'm currently co-writing a manuscript on notation, and using subscripts that don't represent powers is one of our no-no's!
Yes, two sets, so we could subscript for fleet f or something like that if needed, but it might be cleaner to just explain that in the text. The bounds are the same for the two fleets (-5 to 9 as shown in Table 17).
Yes, equation (8) could either have an f subscript on the L or refine text to note that this contribution is for each fleet.
Yes, I didn't know about \mbox but that sounds like an improvement. Happy to have things cleaned up.
Also, I just noticed typos in my new equation (12) so will fix that right away.
Please edit as you wish or instruct me on any changes that you want me to make.
Thanks - I can do the edits. Yes, usually I'd have a fleet subscript, but since we don't have other equations it will be simpler to just state that it occurs twice, and the bounds are the same for the two fleets.
Equation (12) fixed in 26e528d1ab96e11503daf3f040e84220a3ec6aa9
Ian - please take a look through the above commit. I did the changes as specified in 1, 2 and 3 above, plus a few more.
One thing - it's not clear what the value of 7 is referring to in the penultimate sentence.
btw, Ian commented on \mbox above. I've always used that but I think \text is better (seems to scale properly), which is what you've used elsewhere.
Changes all look good to me.
I think the capital \Phi (issue #387) is an improvement, but it needs to be added to equation (8) as well.
This origin of the 7 in this phrase at the end is indeed a bit unclear:
The resulting value of 7 at the mid-point between the bounds is an approximation to the relative impact of a deviation in the logistic parameter space to one in the standard parameter space. The resulting value of Φ = 1.40 approximates the value of φ = 0.20 used in the transformed parameter space in the 2017 assessment.
7 is the derivative of p_{ay} with respect to \tilde{p_{ay}} at \tilde{p_{ay}} = 2, the mid-point between the bounds of -5 and 9. However, since this transformation is applied to each p_a for ages 2 to 6, and those values vary between about 0 and 2.5 as shown in Figure 52(a), the derivative isn't exactly 7 for any age. And the slope is only a linear approximation to the logistic anyway.
Another way to define the 7 would be to say it's the horizontal arrow is about 7 times longer than the vertical arrow in the figure pasted into this comment: https://github.com/cgrandin/hake-assessment/issues/387#issuecomment-364605940.
My struggle with this is to find the balance between being precise about exactly what changed, and not getting into the weeds on a bunch of stuff that in the end really doesn't matter much since the two models line up pretty well.
In hindsight, it was a mistake not to notice the logistic transformation when time-varying selectivity was explored in the MSE in 2013 or 2014, discover that it didn't make any significant difference other than changing the interpretation of phi, and get rid of it back then.
Sorry, I forgot to mention that I already added \mbox to a few other places, such as equation (9), where I do see that the "a_max" looks a bit strange. If you have time to go through and change to \text, please do, or delegate to me to do so.
Good spot about eqn (8). Will do that now.
Can we just omit the sentence about the 7? The important bit is the resulting Phi=1.4. You have the figure in case anyone wants more details.
That \mbox doesn't seem to work in limits of a product (yes, I saw that you did it - saved me trying). Am getting it sorted. Should have been simple.....
I think this is all good now. I just said something like \Phi = 7 \phi yesterday. I should leave it to Ian to confirm and close though.
Looks good.
Clarify this. Kelli's noted that the survey selectivity isn't time-varying, so this should be more explicit. Think about....
Clarify this. Kelli's noted that the survey selectivity isn't time-varying, so this should be more explicit. Think about....
| gharchive/issue | 2018-02-09T17:28:27 | 2025-04-01T06:45:19.908700 | {
"authors": [
"andrew-edwards",
"taylori"
],
"repo": "pacific-hake/hake-assessment",
"url": "https://github.com/pacific-hake/hake-assessment/issues/377",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
627108934 | with the current cluster-template only master shows up
when #96 #98 will get merged you will notice that only the k8s master gets created, the workers do not for some reason.
we use DeploymentMachine for that, and I remember it worked a week ago, so probably something got lost along the way
I see it too. #98 is merged in, and #96, I just cheated by setting the PACKET_API_KEY to base64 encoded, so it all worked.
The master Machine came up, but nothing for the workers based on MachineDeployment.
Nothing in the packet manager logs either, which means it is not even calling them.
I think maybe I found your problem. PR coming soon
No, no PR needed. The default for WORKER_MACHINE_COUNT is 0. If you set it to something else, then it works. You can do it as an env var or as --worker-machine-count when calling clusterctl.
I will update the docs.
| gharchive/issue | 2020-05-29T08:50:25 | 2025-04-01T06:45:19.926997 | {
"authors": [
"deitch",
"gianarb"
],
"repo": "packethost/cluster-api-provider-packet",
"url": "https://github.com/packethost/cluster-api-provider-packet/issues/99",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
492077676 | Creating projects
Implement service.project_create for GitLab.
(reason: I would like to use OGR in my figitool and this functionality is missing.)
LGTM +1
Thanks!
Please, do not merge yet and wait for #190 .
lachmanfrantisek changed the title Creating projects Creating Gitlab projects
I'll add GitHub/Pagure in the next PR.
| gharchive/pull-request | 2019-09-11T07:58:53 | 2025-04-01T06:45:19.929599 | {
"authors": [
"lachmanfrantisek"
],
"repo": "packit-service/ogr",
"url": "https://github.com/packit-service/ogr/pull/189",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
459042045 | Introductory description and how to install GitHub app.
Signed-off-by: Petr "Stone" Hracek phracek@redhat.com
This pull request adds introductory documentation and two approaches on how to install
"Packit-as-a-Service" Github application into projects.
This is the link to the app: https://github.com/apps/packit-as-a-service
Relevant pull request in repository Packit.dev is https://github.com/packit-service/packit.dev/pull/15
Closing this PR. Packit-as-a-Service will be documented in packit.dev repository.
| gharchive/pull-request | 2019-06-21T06:55:50 | 2025-04-01T06:45:19.932212 | {
"authors": [
"TomasTomecek",
"phracek"
],
"repo": "packit-service/packit-service",
"url": "https://github.com/packit-service/packit-service/pull/19",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
630778891 | [files/install-deps(-worker).yaml] install python3-celery & python3-redis
They'd be pip installed anyway, but better be explicit.
Fixes https://sentry.io/organizations/red-hat-0p/issues/1707007212
LGTM
| gharchive/pull-request | 2020-06-04T12:28:14 | 2025-04-01T06:45:19.933591 | {
"authors": [
"jpopelka",
"sakalosj"
],
"repo": "packit-service/packit-service",
"url": "https://github.com/packit-service/packit-service/pull/661",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1551007554 | chore: add pre-commit to README
There was no mention of pre-commit for local development. This adds a
short description of what it does and how to install it
TODO:
[ ] ‹fill in›
Fixes
Related to #206
Merge before/after
RELEASE NOTES BEGIN
Add pre-commit info to README
RELEASE NOTES END
nice, can you add a link to our contributing guidelines? it contains more info & links
Did this, wdyt?
| gharchive/pull-request | 2023-01-20T15:07:55 | 2025-04-01T06:45:19.936636 | {
"authors": [
"SpyTec"
],
"repo": "packit/dashboard",
"url": "https://github.com/packit/dashboard/pull/208",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
853195475 | packit cli file sync should respect git tracked files
When I've used
packit propose-downstream --dist-git-branch epel8 without --local-content it behaves strange:
file sync copy git untracked files
untracked files behaves as --local-content because git did not change version
tracked files are changed to tagged version (it is fine)
possible easy solution could be filter copying just to listed files:
git ls-tree -r --name-only $(git branch --show-current)
From that perspective packit-service and packit cli behaves very differently, because packit-service have everytime proper clean git repo with proper versions.
So that use case that someone workaround packit-service by local packit e.g. in case PS is down may lead to unexpected behaviour.
This is a good one and we discussed it today on our architecture meeting. The conclusion was that @csomh implemented it in the rewritten propose-update command.
Thanks a lot. I'm looking forward for this.
We should make sure that generated spec files are correctly synced downstream.
@csomh Haven't you rewritten the sync logic to rsync? I think this is not relevant anymore, is it?
Tbh, I'm not sure that using rsync as a backend fixed this or not. @jscotka could you please share the steps and description of the state to reproduce this issue, or try it again with the current version of packit?
@csomh @lachmanfrantisek
It is very dependent on what we want to do.
rsync probably fix some issues (at least it can fix the issue, when removing files, if rsync uses the option --delete what will remove extra files on target) (but I can imagine that someone may use it as feature, that has some e.g. extra tests in dowstream and it may break this workflow. so that this --delete should be configurable in packit yaml)
But my fist issue were, that to downstream there were also proposed files what were not added to git. (e.g. I've created some temp file) it although packit checkout some release tag, these files are still there and present and was synced as well.
So there is more possible solutions (my proposed were, to just have there filter and copy just git tracked files (or maybe copy current dir to some temp one, do git reset and then do all packit stuff))
| gharchive/issue | 2021-04-08T08:08:05 | 2025-04-01T06:45:19.942556 | {
"authors": [
"TomasTomecek",
"csomh",
"jscotka",
"lachmanfrantisek"
],
"repo": "packit/packit",
"url": "https://github.com/packit/packit/issues/1198",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1092133996 | Failed to install shfmt-bin
Describe the bug
/usr/share/pacstall/scripts/install-local.sh: line 278: shfmt-bin.pacscript: No such file or directory
pacscript not found.
To Reproduce
Steps to reproduce the behavior:
Ran pacstall -PI shfmt-bin.
The pacscript doesn't exist.
Expected behavior
I expected the program to install properly.
Screenshots
Desktop (please complete the following information):
OS: [Ubuntu WSL]
Version [20.04.3 LTS on Windows 10 x86_64]
Can you verify the pacscript by choosing yes when prompted for editing/viewing the script?
Also try running sudo chown -R $USER:$USER /tmp/pacstall && sudo chown -R $USER:$USER /var/log/pacstall
| gharchive/issue | 2022-01-03T00:17:30 | 2025-04-01T06:45:19.959606 | {
"authors": [
"Henryws",
"pure-cheekbones"
],
"repo": "pacstall/pacstall-programs",
"url": "https://github.com/pacstall/pacstall-programs/issues/610",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1067500528 | feat(precommit): Add it
Fixes #536
Waiting on resolution of https://github.com/mvdan/sh/issues/715
| gharchive/pull-request | 2021-11-30T17:33:13 | 2025-04-01T06:45:19.960753 | {
"authors": [
"wizard-28"
],
"repo": "pacstall/pacstall-programs",
"url": "https://github.com/pacstall/pacstall-programs/pull/537",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
196219195 | Error charset into ID Markdown element
Problem
Incorrect output h2, h3 where generate ID for this elements
Technical Details
Pagekit version: 1.0.9
Hi Kamuz,
The Markdown ID parser is not the best, I already reported a fix for another bug https://github.com/pagekit/pagekit/pull/751 , I feel that the IDs parser itself needs more detailed revision.
Thanks,
Grigor
| gharchive/issue | 2016-12-17T12:16:04 | 2025-04-01T06:45:20.019530 | {
"authors": [
"grigoryosifov",
"kamuz"
],
"repo": "pagekit/pagekit",
"url": "https://github.com/pagekit/pagekit/issues/785",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
1990402182 | feat: Update STANDIN metadata [PAGOPA-1348]
List of Changes
Move STANDIN metadata conditional creation from PaymentOption to Transfer.
Update Unit test.
Motivation and Context
How Has This Been Tested?
Unit test
Test in local environment
Screenshots (if appropriate):
Types of changes
[ ] Bug fix (non-breaking change which fixes an issue)
[ ] New feature (non-breaking change which adds functionality)
[ ] Breaking change (fix or feature that would cause existing functionality to not work as expected)
Checklist:
[ ] My change requires a change to the documentation.
[x] I have updated the documentation accordingly.
This issue cannot be handled with a metadata because it could be changed during the history of a debt position and payment.
| gharchive/pull-request | 2023-11-13T10:54:19 | 2025-04-01T06:45:20.042410 | {
"authors": [
"AngeloCaporaso"
],
"repo": "pagopa/pagopa-debt-position",
"url": "https://github.com/pagopa/pagopa-debt-position/pull/193",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2110146549 | No cooling circuit values WPE-I 06 HKW 230 Premium
Source values seem to be a firmware error because I can't get them manually via Modbus
But the others work manually:
- name: WP_WP_Rücklauf
unit_of_measurement: °C
device_class: temperature
address: 541
scale: 0.1
slave: 1
input_type: input
data_type: int16
- name: WP_WP_Heissgas
unit_of_measurement: °C
device_class: temperature
address: 543
scale: 0.1
slave: 1
input_type: input
data_type: int16
- name: WP_WP_Niederdruck
unit_of_measurement: bar
device_class: pressure
address: 544
scale: 0.01
slave: 1
input_type: input
data_type: int16
- name: WP_WP_Hochdruck
unit_of_measurement: bar
device_class: pressure
address: 546
scale: 0.01
slave: 1
input_type: input
data_type: int16
Diagnostic Infos: config_entry-stiebel_eltron_isg-717248ecded5b0d28c57a477522b016f.json (1).txt
Sorry, can you please be more precise on what exactly you are experiencing?
Do you need update the firmware of the ISG?
Hello,
I think the problem is that the source values (source temperature and source pressure) are not correctly coming from the isg. There stiebel Eltron has to make an firmware update.
But it is possible to get the other values from the screenshot. That ist what his code is showing. But in the application they were not showing. At my stiebel Eltron wpl-a 07 it is still the same. I got no hot gas value or the others of the screenshot at the isg entity.
The first part of your problem is possibly related to https://github.com/pail23/stiebel_eltron_isg_component/issues/176
If this is so, it can't be fixed at this time.
The second part, still I don't understand what you are saying. Can you please be more specific, what you are expecting and what do you see.
Hi,
That's right, the first part is in #176. I hadn't found it before. I also asked Stiebel at the same time.
The other problem is the lack of hot gas and low pressure values.
This is because some heat pumps use different addresses.
I have adapted the files and am trying to create a pull.
Tested with the heat pumps: Stiebel Eltron WPE-I XX XXX 230 Premium and Stiebel Eltron WPL-A
Hello,
every entity at the screenshot he postet with „source“ in the name is a drivel problem.
but the others at the screenshot were shown as not available but the way the thread starter postet it, they will be shown with correct values.
| gharchive/issue | 2024-01-31T13:53:42 | 2025-04-01T06:45:20.054609 | {
"authors": [
"MiJo112",
"Padanian",
"Schatti93"
],
"repo": "pail23/stiebel_eltron_isg_component",
"url": "https://github.com/pail23/stiebel_eltron_isg_component/issues/184",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1590240885 | build image un-idempotency
recently,I found that build image almost failed eachtime, and some buildpack‘behavior
not so consistency(eg: they may 'toggle' passed or not when re-run build image)
Expected Behavior
build a image successfully or tell me why failed with more detail
Current Behavior
build image failed
Possible Solution
Steps to Reproduce
here is my shell script and you can find all information in my repo https://github.com/TianMing2018/paketo-demo#issues
# pack war
mvn clean package
# run three times and save log to nbg-pack*.log
sudo pack build pack-demo \
--verbose \
--buildpack paketobuildpacks/java:8 \
--buildpack paketobuildpacks/health-checker:1 \
--builder paketobuildpacks/builder:base \
--volume proxy-volume:/platform/bindings/dependency-mapping:ro \
--pull-policy if-not-present \
--env BP_LOG_LEVEL=DEBUG \
--env BP_JVM_TYPE=JDK \
--env BP_JVM_VERSION=11 \
--env BP_HEALTH_CHECKER_ENABLED=true \
--env THC_PORT=8180 \
--env THC_PATH="/actuator/health" \
--env BPE_LANG=zh_CN.UTF-8 \
--env BPE_LC_ALL=zh_CN.UTF-8 \
--env LANG=zh_CN.UTF-8 \
--env LC_ALL=zh_CN.UTF-8 \
--path target/paketo-demo-0.0.1-SNAPSHOT.war > nbg-pack1.log
# run three times and save log to nbg-sb*.log
mvn install -P paketo -l nbg-sb1.log
Motivations
build a image with health-check ability
It seems that the builder crushed when it try to bind volume to buildpack ,cause of there was multi-version of cached jdk and syft in the volume, here is the files list in my proxy-volume,when I remove lower(or discarded) version ,I can build image successfully everytime。I'm wondering what you will do to solve this issue ? add some description in docs guides / add validation in buildpack / or just do nothing ,please leave a comment.
#### file in proxy-volume
root@ac4d547e9e30:/# ls -al
total 439080
drwxr-xr-x 2 root root 4096 Feb 18 11:38 .
drwxrwxrwt 1 root root 4096 Feb 18 11:41 ..
-rw-r--r-- 1 root root 87 Jan 29 05:57 2ea2f787065a1ac58b528a4d7db453fa0f8e05b2192300b4ea52ecf885723137
-rw-r--r-- 1 root root 75 Feb 18 11:11 30912a77b1b4a565cf6ed936ea619aa353d8d9ddc5d216afdc242cf3b1e7a0a7
-rw-r--r-- 1 root root 76 Feb 18 11:11 71296b3d87b2837aeb7e55b2ca98dc69f92c7274ae3d968698da5887b7462ad2
-rw-r--r-- 1 root root 86 Dec 17 08:27 86df083bd817970404d56d360926d440eb895041ab8d7f580d79b34b19e621e8
-rw-r--r-- 1 root root 76 Dec 17 08:33 89860504694a05a75688991ac24281cb84cfa61d48c973ddee7559fa7fc0a60e
-rw-r--r-- 1 root root 78 Jan 29 05:57 9f76c9c93748472a9f145999d4da99c1d8cb070872bba574843c45ef1d14c27f
-rw-r--r-- 1 root root 208333155 Dec 12 16:40 bellsoft-jdk11.0.17+7-linux-amd64.tar.gz
-rw-r--r-- 1 1000 1000 208389475 Jan 16 14:10 bellsoft-jdk11.0.18+10-linux-amd64.tar.gz
-rw-r--r-- 1 root root 75 Dec 19 11:29 eca042851d6dc8c2158a923aaf0736d8b9141a10ee8b77a2a45c9528ef74799b
-rw-r--r-- 1 1000 1000 61385 Jan 10 12:06 spring-cloud-bindings-1.11.0.jar
-rw-r--r-- 1 root root 20014858 Dec 14 01:40 syft_0.62.1_linux_amd64.tar.gz
-rw-r--r-- 1 1000 1000 10878369 Jan 31 17:13 syft_0.69.1_linux_amd64.tar.gz
-rw-r--r-- 1 1000 1000 837832 Jan 31 04:51 thc-x86_64-unknown-linux-musl
-rw-r--r-- 1 1000 1000 1050824 Jan 9 04:11 thc-x86_64-unknown-linux-musl.bak
-rw-r--r-- 1 root root 19 Dec 17 08:18 type
#### file content
root@ac4d547e9e30:/# cat 2ea2f787065a1ac58b528a4d7db453fa0f8e05b2192300b4ea52ecf885723137
file:///platform/bindings/dependency-mapping/bellsoft-jdk11.0.18+10-linux-amd64.tar.gz
root@ac4d547e9e30:/# cat 30912a77b1b4a565cf6ed936ea619aa353d8d9ddc5d216afdc242cf3b1e7a0a7
file:///platform/bindings/dependency-mapping/thc-x86_64-unknown-linux-musl
root@ac4d547e9e30:/# cat 71296b3d87b2837aeb7e55b2ca98dc69f92c7274ae3d968698da5887b7462ad2
file:///platform/bindings/dependency-mapping/syft_0.69.1_linux_amd64.tar.gz
root@ac4d547e9e30:/# cat 86df083bd817970404d56d360926d440eb895041ab8d7f580d79b34b19e621e8
file:///platform/bindings/dependency-mapping/bellsoft-jdk11.0.17+7-linux-amd64.tar.gz
root@ac4d547e9e30:/# cat 89860504694a05a75688991ac24281cb84cfa61d48c973ddee7559fa7fc0a60e
file:///platform/bindings/dependency-mapping/syft_0.62.1_linux_amd64.tar.gz
root@ac4d547e9e30:/# cat 9f76c9c93748472a9f145999d4da99c1d8cb070872bba574843c45ef1d14c27f
file:///platform/bindings/dependency-mapping/spring-cloud-bindings-1.11.0.jar
root@ac4d547e9e30:/# cat eca042851d6dc8c2158a923aaf0736d8b9141a10ee8b77a2a45c9528ef74799b
file:///platform/bindings/dependency-mapping/thc-x86_64-unknown-linux-musl
#### remove outdated file
root@ac4d547e9e30:/# rm eca042851d6dc8c2158a923aaf0736d8b9141a10ee8b77a2a45c9528ef74799b
root@ac4d547e9e30:/# rm thc-x86_64-unknown-linux-musl.bak
root@ac4d547e9e30:/# rm 89860504694a05a75688991ac24281cb84cfa61d48c973ddee7559fa7fc0a60e
root@ac4d547e9e30:/# rm syft_0.62.1_linux_amd64.tar.gz
root@ac4d547e9e30:/# rm 86df083bd817970404d56d360926d440eb895041ab8d7f580d79b34b19e621e8
root@ac4d547e9e30:/# rm bellsoft-jdk11.0.17+7-linux-amd64.tar.gz
Is this still an issue? I'm not seeing a specific error here? When do you see a failure and what is the exact error message you see?
Is this still an issue? I'm not seeing a specific error here? When do you see a failure and what is the exact error message you see?
I had describe the problem below,Shortly, when we use volume to store multi jdk and others files, the build will be messed, it's difficult to tell the problem, maybe we should quote this point in documents or modify code
Is there any way you could attach a build log that shows what happens? I don't think I understand what you're reporting here.
I can see you're hitting some sort of issue with mapped dependencies and in particular when there are multiple versions, but I have had this case before. I frequently use mapped dependencies and haven't encountered this issue so I definitely want to hear more about it and try to understand what might be happening. I just need a little more info to go on, build logs showing the errors you see (or what you don't see, depending on the problem), and more concise steps to reproduce. I looked at your linked project, but I'm not exactly sure what to do with it to reproduce the problem. If you could give me a brief set of steps to reproduce, I could try that also. Thanks
Sorry,I didn't persist log files at that moment, and I re-run test now but can't reproduction issue, So I decide to close this issue for now.
| gharchive/issue | 2023-02-18T07:10:05 | 2025-04-01T06:45:20.061855 | {
"authors": [
"TianMing2018",
"dmikusa"
],
"repo": "paketo-buildpacks/java",
"url": "https://github.com/paketo-buildpacks/java/issues/914",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
684686271 | Add NGINX Buildpack documentation
We should add public facing documentation on the website for the NGINX Paketo Buildpack.
This should be under the new docs navbar (see https://github.com/paketo-buildpacks/paketo-website/issues/24) as a new page called Paketo NGINX Buildpack.
These docs should be application developer facing and should at the very least document things like:
Possible app configurations supported by the buildpack
Buildpackage artifacts on GCR and how to build a NGINX app using a buildpackage (e.g pack build --buildpack gcr.io/paketo-buildpacks/nginx)
Supported versions of dependencies (link to release notes)
Buildpack.yml configurations (specifying specific NGINX version, etc.)
vendoring dependencies (for offline use case)
buildpack specific environment variables
This looks great. This has been implemented in the redesign branch and will be merged in once we're closer to shipping all of our website changes.
| gharchive/issue | 2020-08-24T13:47:45 | 2025-04-01T06:45:20.065703 | {
"authors": [
"kvedurmu"
],
"repo": "paketo-buildpacks/paketo-website",
"url": "https://github.com/paketo-buildpacks/paketo-website/issues/32",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2445650036 | Update buildpack.toml to latest yarn
Summary
Update the ruby buildpack to pull in the latest yarn buildpack to fix the installation of yarn
Use Cases
Checklist
[x] I have viewed, signed, and submitted the Contributor License Agreement.
[ ] I have linked issue(s) that this PR should close using keywords or the Github UI (See docs)
[x] I have added an integration test, if necessary.
[x ] I have reviewed the styleguide for guidance on my code quality.
[ x] I'm happy with the commit history on this PR (I have rebased/squashed as needed).
https://github.com/paketo-buildpacks/ruby/releases/tag/v0.47.2
| gharchive/pull-request | 2024-08-02T19:37:36 | 2025-04-01T06:45:20.069199 | {
"authors": [
"a1flecke",
"sophiewigmore"
],
"repo": "paketo-buildpacks/ruby",
"url": "https://github.com/paketo-buildpacks/ruby/pull/1143",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1917789153 | more .syncignore fixup
Summary
avoid addition of receipts.sh since not being used for ubi stack
Use Cases
Checklist
[x] I have viewed, signed, and submitted the Contributor License Agreement.
[x] I have linked issue(s) that this PR should close using keywords or the Github UI (See docs)
[x] I have added an integration test, if necessary.
[x] I have reviewed the styleguide for guidance on my code quality.
[x] I'm happy with the commit history on this PR (I have rebased/squashed as needed).
I think https://github.com/paketo-community/ubi-base-stack/pull/14 needs to land for the CI in this PR to pass.
added these changes to the github-config PR
| gharchive/pull-request | 2023-09-28T15:29:22 | 2025-04-01T06:45:20.073295 | {
"authors": [
"mhdawson",
"sophiewigmore"
],
"repo": "paketo-community/ubi-base-stack",
"url": "https://github.com/paketo-community/ubi-base-stack/pull/15",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
199581413 | jsx-curly-spacing has lint issues
There are four const related lint issues in jsx-curly-spacing:
src/rules/jsxCurlySpacingRule.ts[130, 21]: Identifier 'failureString' is never reassigned; use 'const' instead of 'let'.
src/rules/jsxCurlySpacingRule.ts[140, 21]: Identifier 'failureString' is never reassigned; use 'const' instead of 'let'.
src/rules/jsxCurlySpacingRule.ts[151, 25]: Identifier 'failureString' is never reassigned; use 'const' instead of 'let'.
src/rules/jsxCurlySpacingRule.ts[160, 25]: Identifier 'failureString' is never reassigned; use 'const' instead of 'let'.
yep, this happens sometimes when we update the tslint:latest config. thanks for the PR.
| gharchive/issue | 2017-01-09T15:35:20 | 2025-04-01T06:45:20.109604 | {
"authors": [
"adidahiya",
"no23reason"
],
"repo": "palantir/tslint-react",
"url": "https://github.com/palantir/tslint-react/issues/58",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
114855097 | Make src/tslint.ts a commonjs-style module instead of an ES6 module.
Reverts 'tslint' back to being an old-style module so it works better with CommonJS.
Should fix #760
Downsides: We now have a CommonJS-style module scattered among ES6-style modules
| gharchive/pull-request | 2015-11-03T16:30:47 | 2025-04-01T06:45:20.110865 | {
"authors": [
"JKillian"
],
"repo": "palantir/tslint",
"url": "https://github.com/palantir/tslint/pull/764",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
518488218 | DateField for mongoengine absent in admin
prerequisites:
Flask-Admin 1.5.4
mongoengine 0.18.2
for example
class User(Document):
last_active_action = DateField(required=True)
# admin.py
class UserView(ModelView):
pass
admin.add_view(
UserView(User, name="User", endpoint="admin-user", menu_class_name="User")
)
it will be crash, because this field will be no show
I am closing this issue as we're removing support for flask-mongoengine in the next release, because that library is no longer being maintained and is incompatible with recent versions of our core dependencies (eg flask, wtforms, etc).
| gharchive/issue | 2019-11-06T13:56:07 | 2025-04-01T06:45:20.116548 | {
"authors": [
"malinich",
"samuelhwilliams"
],
"repo": "pallets-eco/flask-admin",
"url": "https://github.com/pallets-eco/flask-admin/issues/1921",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
784120344 | Use of depreciated & insecure tempfile.mktemp()
Expected Behavior
_tempfilepager() in _termui_impl.py uses a depreciated function that is considered insecure by the core Python team. It is recommended to use mkstemp() instead.
import tempfile
filename = tempfile.mkstemp() # Suggested replacement
Actual Behavior
Python's warning:
Use of this function may introduce a security hole in your program. By the time you get around to doing anything with the file name it returns, someone else may have beaten you to the punch.
import tempfile
filename = tempfile.mktemp() # Depreciated and insecure
Environment
Python version: 3.8.5
Click version: 7.1.2
Please see our guidelines for responsibly reporting security issues: https://github.com/pallets/click/security/policy
Please see our guidelines for responsibly reporting security issues: https://github.com/pallets/click/security/policy
| gharchive/issue | 2021-01-12T10:42:59 | 2025-04-01T06:45:20.120397 | {
"authors": [
"MIS-Alex",
"davidism"
],
"repo": "pallets/click",
"url": "https://github.com/pallets/click/issues/1752",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
177647803 | Removed unused import.
This change is
Thanks!
| gharchive/pull-request | 2016-09-18T12:00:10 | 2025-04-01T06:45:20.121778 | {
"authors": [
"Thalmann",
"untitaker"
],
"repo": "pallets/flask",
"url": "https://github.com/pallets/flask/pull/2026",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
2290785516 | tests/test_leak.py::test_markup_leaks fails with JIT-enabled Python 3.13.0b1
When Python 3.13.0b1 is built with --enable-experimental-jit, I'm getting the following test failure:
$ tox -e py313
.pkg-cpython313: _optional_hooks> python /usr/lib/python3.12/site-packages/pyproject_api/_backend.py True setuptools.build_meta
.pkg-cpython313: get_requires_for_build_wheel> python /usr/lib/python3.12/site-packages/pyproject_api/_backend.py True setuptools.build_meta
.pkg-cpython313: build_wheel> python /usr/lib/python3.12/site-packages/pyproject_api/_backend.py True setuptools.build_meta
py313: install_package> python -I -m pip install --force-reinstall --no-deps /tmp/markupsafe/.tox/.tmp/package/2/MarkupSafe-3.0.0.dev0-cp313-cp313-linux_x86_64.whl
py313: commands[0]> pytest -v --tb=short --basetemp=/tmp/markupsafe/.tox/py313/tmp
========================================================= test session starts =========================================================
platform linux -- Python 3.13.0b1, pytest-8.2.0, pluggy-1.5.0 -- /tmp/markupsafe/.tox/py313/bin/python
cachedir: .tox/py313/.pytest_cache
rootdir: /tmp/markupsafe
configfile: pyproject.toml
testpaths: tests
collected 74 items
tests/test_escape.py::test_escape[markupsafe._native--] PASSED [ 1%]
tests/test_escape.py::test_escape[markupsafe._native-abcd&><'"efgh-abcd&><'"efgh] PASSED [ 2%]
tests/test_escape.py::test_escape[markupsafe._native-&><'"efgh-&><'"efgh] PASSED [ 4%]
tests/test_escape.py::test_escape[markupsafe._native-abcd&><'"-abcd&><'"] PASSED [ 5%]
tests/test_escape.py::test_escape[markupsafe._native-\u3053\u3093\u306b\u3061\u306f&><'"\u3053\u3093\u3070\u3093\u306f-\u3053\u3093\u306b\u3061\u306f&><'"\u3053\u3093\u3070\u3093\u306f] PASSED [ 6%]
tests/test_escape.py::test_escape[markupsafe._native-&><'"\u3053\u3093\u3070\u3093\u306f-&><'"\u3053\u3093\u3070\u3093\u306f] PASSED [ 8%]
tests/test_escape.py::test_escape[markupsafe._native-\u3053\u3093\u306b\u3061\u306f&><'"-\u3053\u3093\u306b\u3061\u306f&><'"] PASSED [ 9%]
tests/test_escape.py::test_escape[markupsafe._native-\U0001f363\U0001f362&><'"\U0001f37a xyz-\U0001f363\U0001f362&><'"\U0001f37a xyz] PASSED [ 10%]
tests/test_escape.py::test_escape[markupsafe._native-&><'"\U0001f37a xyz-&><'"\U0001f37a xyz] PASSED [ 12%]
tests/test_escape.py::test_escape[markupsafe._native-\U0001f363\U0001f362&><'"-\U0001f363\U0001f362&><'"] PASSED [ 13%]
tests/test_exception_custom_html.py::test_exception_custom_html[markupsafe._native] PASSED [ 14%]
tests/test_leak.py::test_markup_leaks[markupsafe._native] FAILED [ 16%]
tests/test_markupsafe.py::test_adding[markupsafe._native] PASSED [ 17%]
tests/test_markupsafe.py::test_string_interpolation[markupsafe._native-<em>%s</em>-<bad user>-<em><bad user></em>] PASSED [ 18%]
tests/test_markupsafe.py::test_string_interpolation[markupsafe._native-<em>%(username)s</em>-data1-<em><bad user></em>] PASSED [ 20%]
tests/test_markupsafe.py::test_string_interpolation[markupsafe._native-%i-3.14-3] PASSED [ 21%]
tests/test_markupsafe.py::test_string_interpolation[markupsafe._native-%.2f-3.14-3.14] PASSED [ 22%]
tests/test_markupsafe.py::test_type_behavior[markupsafe._native] PASSED [ 24%]
tests/test_markupsafe.py::test_html_interop[markupsafe._native] PASSED [ 25%]
tests/test_markupsafe.py::test_missing_interpol[markupsafe._native-foo] PASSED [ 27%]
tests/test_markupsafe.py::test_missing_interpol[markupsafe._native-42] PASSED [ 28%]
tests/test_markupsafe.py::test_missing_interpol[markupsafe._native-args2] PASSED [ 29%]
tests/test_markupsafe.py::test_tuple_interpol[markupsafe._native] PASSED [ 31%]
tests/test_markupsafe.py::test_dict_interpol[markupsafe._native] PASSED [ 32%]
tests/test_markupsafe.py::test_escaping[markupsafe._native] PASSED [ 33%]
tests/test_markupsafe.py::test_unescape[markupsafe._native] PASSED [ 35%]
tests/test_markupsafe.py::test_format[markupsafe._native] PASSED [ 36%]
tests/test_markupsafe.py::test_format_map[markupsafe._native] PASSED [ 37%]
tests/test_markupsafe.py::test_formatting_empty[markupsafe._native] PASSED [ 39%]
tests/test_markupsafe.py::test_custom_formatting[markupsafe._native] PASSED [ 40%]
tests/test_markupsafe.py::test_complex_custom_formatting[markupsafe._native] PASSED [ 41%]
tests/test_markupsafe.py::test_formatting_with_objects[markupsafe._native] PASSED [ 43%]
tests/test_markupsafe.py::test_escape_silent[markupsafe._native] PASSED [ 44%]
tests/test_markupsafe.py::test_splitting[markupsafe._native] PASSED [ 45%]
tests/test_markupsafe.py::test_mul[markupsafe._native] PASSED [ 47%]
tests/test_markupsafe.py::test_escape_return_type[markupsafe._native] PASSED [ 48%]
tests/test_markupsafe.py::test_soft_str[markupsafe._native] PASSED [ 50%]
tests/test_escape.py::test_escape[markupsafe._speedups--] PASSED [ 51%]
tests/test_escape.py::test_escape[markupsafe._speedups-abcd&><'"efgh-abcd&><'"efgh] PASSED [ 52%]
tests/test_escape.py::test_escape[markupsafe._speedups-&><'"efgh-&><'"efgh] PASSED [ 54%]
tests/test_escape.py::test_escape[markupsafe._speedups-abcd&><'"-abcd&><'"] PASSED [ 55%]
tests/test_escape.py::test_escape[markupsafe._speedups-\u3053\u3093\u306b\u3061\u306f&><'"\u3053\u3093\u3070\u3093\u306f-\u3053\u3093\u306b\u3061\u306f&><'"\u3053\u3093\u3070\u3093\u306f] PASSED [ 56%]
tests/test_escape.py::test_escape[markupsafe._speedups-&><'"\u3053\u3093\u3070\u3093\u306f-&><'"\u3053\u3093\u3070\u3093\u306f] PASSED [ 58%]
tests/test_escape.py::test_escape[markupsafe._speedups-\u3053\u3093\u306b\u3061\u306f&><'"-\u3053\u3093\u306b\u3061\u306f&><'"] PASSED [ 59%]
tests/test_escape.py::test_escape[markupsafe._speedups-\U0001f363\U0001f362&><'"\U0001f37a xyz-\U0001f363\U0001f362&><'"\U0001f37a xyz] PASSED [ 60%]
tests/test_escape.py::test_escape[markupsafe._speedups-&><'"\U0001f37a xyz-&><'"\U0001f37a xyz] PASSED [ 62%]
tests/test_escape.py::test_escape[markupsafe._speedups-\U0001f363\U0001f362&><'"-\U0001f363\U0001f362&><'"] PASSED [ 63%]
tests/test_exception_custom_html.py::test_exception_custom_html[markupsafe._speedups] PASSED [ 64%]
tests/test_leak.py::test_markup_leaks[markupsafe._speedups] FAILED [ 66%]
tests/test_markupsafe.py::test_adding[markupsafe._speedups] PASSED [ 67%]
tests/test_markupsafe.py::test_string_interpolation[markupsafe._speedups-<em>%s</em>-<bad user>-<em><bad user></em>] PASSED [ 68%]
tests/test_markupsafe.py::test_string_interpolation[markupsafe._speedups-<em>%(username)s</em>-data1-<em><bad user></em>] PASSED [ 70%]
tests/test_markupsafe.py::test_string_interpolation[markupsafe._speedups-%i-3.14-3] PASSED [ 71%]
tests/test_markupsafe.py::test_string_interpolation[markupsafe._speedups-%.2f-3.14-3.14] PASSED [ 72%]
tests/test_markupsafe.py::test_type_behavior[markupsafe._speedups] PASSED [ 74%]
tests/test_markupsafe.py::test_html_interop[markupsafe._speedups] PASSED [ 75%]
tests/test_markupsafe.py::test_missing_interpol[markupsafe._speedups-foo] PASSED [ 77%]
tests/test_markupsafe.py::test_missing_interpol[markupsafe._speedups-42] PASSED [ 78%]
tests/test_markupsafe.py::test_missing_interpol[markupsafe._speedups-args2] PASSED [ 79%]
tests/test_markupsafe.py::test_tuple_interpol[markupsafe._speedups] PASSED [ 81%]
tests/test_markupsafe.py::test_dict_interpol[markupsafe._speedups] PASSED [ 82%]
tests/test_markupsafe.py::test_escaping[markupsafe._speedups] PASSED [ 83%]
tests/test_markupsafe.py::test_unescape[markupsafe._speedups] PASSED [ 85%]
tests/test_markupsafe.py::test_format[markupsafe._speedups] PASSED [ 86%]
tests/test_markupsafe.py::test_format_map[markupsafe._speedups] PASSED [ 87%]
tests/test_markupsafe.py::test_formatting_empty[markupsafe._speedups] PASSED [ 89%]
tests/test_markupsafe.py::test_custom_formatting[markupsafe._speedups] PASSED [ 90%]
tests/test_markupsafe.py::test_complex_custom_formatting[markupsafe._speedups] PASSED [ 91%]
tests/test_markupsafe.py::test_formatting_with_objects[markupsafe._speedups] PASSED [ 93%]
tests/test_markupsafe.py::test_escape_silent[markupsafe._speedups] PASSED [ 94%]
tests/test_markupsafe.py::test_splitting[markupsafe._speedups] PASSED [ 95%]
tests/test_markupsafe.py::test_mul[markupsafe._speedups] PASSED [ 97%]
tests/test_markupsafe.py::test_escape_return_type[markupsafe._speedups] PASSED [ 98%]
tests/test_markupsafe.py::test_soft_str[markupsafe._speedups] PASSED [100%]
============================================================== FAILURES ===============================================================
________________________________________________ test_markup_leaks[markupsafe._native] ________________________________________________
tests/test_leak.py:31: in test_markup_leaks
assert len(counts) == 1
E assert 2 == 1
E + where 2 = len({31213, 31214})
_______________________________________________ test_markup_leaks[markupsafe._speedups] _______________________________________________
tests/test_leak.py:31: in test_markup_leaks
assert len(counts) == 1
E assert 2 == 1
E + where 2 = len({31954, 31955})
======================================================= short test summary info =======================================================
FAILED tests/test_leak.py::test_markup_leaks[markupsafe._native] - assert 2 == 1
FAILED tests/test_leak.py::test_markup_leaks[markupsafe._speedups] - assert 2 == 1
==================================================== 2 failed, 72 passed in 0.35s =====================================================
py313: exit 1 (0.61 seconds) /tmp/markupsafe> pytest -v --tb=short --basetemp=/tmp/markupsafe/.tox/py313/tmp pid=842994
py313: FAIL code 1 (2.36=setup[1.75]+cmd[0.61] seconds)
evaluation failed :( (2.78 seconds)
Curiously, enabling the gc.collect() call doesn't seem to help here.
Environment:
Gentoo Linux amd64
Python version: 3.13.0b1
MarkupSafe version: 52d35f57ee0ba9dd8d74032f7741c27f27b05edf
Appears to be passing on main. I'm on Mac and did the following:
$ ./configure --disable-gil --with-openssl="$(brew --prefix openssl)" && make -j
$ virtualenv -p ~/Python-3.13.0b1/python.exe .venv313
$ . .venv313/bin/activate.fish
$ pip install -r requirements/dev.txt && pip install -e .
$ PYTHON_GIL=0 pytest
However, if I take out the if pypy condition, and always call gc.collect(), then it does fail on every cpython version (not just 3.13).
Frankly, I do not understand this test (and I did not write it, and there's no comments explaining it). Why do we get a different count after GC, except on PyPy where we get a different count without GC? Why do we only get 2 different counts over 20x1000 loops? Why are we treating this second count as a memory leak in MarkupSafe, it seems like it's just the behavior of Python ref counts and GC at work. I would expect a memory leak to result in a different count most times, not only 2 different counts.
Appears to be passing on main. I'm on Mac and did the following:
$ ./configure --disable-gil --with-openssl="$(brew --prefix openssl)" && make -j
You've missed --enable-experimental-jit, though I don't know if it's support on Mac.
Ah, I missed that it was the JIT and not the GIL. Anyway, my confusion over the entire test still stands.
JIT doesn't build on Mac yet.
Moving gc.collect() to before the loops rather than during them works for cpython and PyPy, but not 3.13 JIT. I think I'm just going to say assert count < 3. I'm guessing some internals stabilize over time resulting in different counts. If this test becomes flaky again in the future I might just remove it.
Closed by #446
Thanks! I can confirm that they pass for me now.
| gharchive/issue | 2024-05-11T08:37:09 | 2025-04-01T06:45:20.132133 | {
"authors": [
"davidism",
"mgorny"
],
"repo": "pallets/markupsafe",
"url": "https://github.com/pallets/markupsafe/issues/445",
"license": "bsd-3-clause",
"license_type": "permissive",
"license_source": "bigquery"
} |
1755542111 | float_format in .to_string can be str
Describe the bug
The change in #721 broke this for mypy. Although the current docs say "one-parameter function, optional, default None", you are allowed to pass a string that specifies the format like in .to_csv.
To Reproduce
These usages all work with both frame and series.
# t.py
import pandas as pd
df = pd.DataFrame({"a": [0.1, 1.0, 3.14]})
print(df.to_string(float_format="%.3g"))
print(df.to_string(float_format="{:.3g}".format))
print(df.to_string(float_format=lambda x: f"{x:.3e}"))
print(df.to_string(float_format=pd.io.formats.format.EngFormatter(3, True)))
print(df.a.to_string(float_format="%.3g"))
print(df.a.to_string(float_format="{:.3g}".format))
print(df.a.to_string(float_format=lambda x: f"{x:.3e}"))
print(df.a.to_string(float_format=pd.io.formats.format.EngFormatter(3, True)))
> mypy t.py
t.py:5: error: No overload variant of "to_string" of "DataFrame" matches argument type "str" [call-overload]
t.py:5: note: Possible overload variants:
t.py:5: note: def [HashableT1 <: Hashable, HashableT2 <: Hashable] to_string(self, buf: Union[str, PathLike[str], WriteBuffer[str]], columns: Union[List[HashableT1], Index, Series[Any], None] = ..., col_space: Union[int, List[int], Dict[HashableT2, int], None] = ..., header: Union[bool, List[str], Tuple[str, ...]] = ..., index: bool = ..., na_rep: str = ..., formatters: Union[List[Callable[..., Any]], Tuple[Callable[..., Any], ...], Mapping[Union[str, int], Callable[..., Any]], None] = ..., float_format: Optional[Callable[[float], str]] = ..., sparsify: Optional[bool] = ..., index_names: bool = ..., justify: Optional[str] = ..., max_rows: Optional[int] = ..., max_cols: Optional[int] = ..., show_dimensions: bool = ..., decimal: str = ..., line_width: Optional[int] = ..., min_rows: Optional[int] = ..., max_colwidth: Optional[int] = ..., encoding: Optional[str] = ...) -> None
t.py:5: note: def [HashableT <: Hashable] to_string(self, buf: None = ..., columns: Union[List[HashableT], Index, Series[Any], None] = ..., col_space: Union[int, List[int], Dict[Hashable, int], None] = ..., header: Union[bool, Sequence[str]] = ..., index: bool = ..., na_rep: str = ..., formatters: Union[List[Callable[..., Any]], Tuple[Callable[..., Any], ...], Mapping[Union[str, int], Callable[..., Any]], None] = ..., float_format: Optional[Callable[[float], str]] = ..., sparsify: Optional[bool] = ..., index_names: bool = ..., justify: Optional[str] = ..., max_rows: Optional[int] = ..., max_cols: Optional[int] = ..., show_dimensions: bool = ..., decimal: str = ..., line_width: Optional[int] = ..., min_rows: Optional[int] = ..., max_colwidth: Optional[int] = ..., encoding: Optional[str] = ...) -> str
t.py:10: error: No overload variant of "to_string" of "Series" matches argument type "str" [call-overload]
t.py:10: note: Possible overload variants:
t.py:10: note: def to_string(self, buf: Union[str, PathLike[str], WriteBuffer[str]], na_rep: str = ..., float_format: Callable[[float], str] = ..., header: bool = ..., index: bool = ..., length: bool = ..., dtype: bool = ..., name: bool = ..., max_rows: Optional[int] = ..., min_rows: Optional[int] = ...) -> None
t.py:10: note: def to_string(self, buf: None = ..., na_rep: str = ..., float_format: Callable[[float], str] = ..., header: bool = ..., index: bool = ..., length: bool = ..., dtype: bool = ..., name: bool = ..., max_rows: Optional[int] = ..., min_rows: Optional[int] = ...) -> str
Found 2 errors in 1 file (checked 1 source file)
Please complete the following information:
OS: [e.g. Windows, Linux, MacOS] Windows
OS Version [e.g. 22] 10
python version 3.10.11
version of type checker mypy 1.3.0
version of installed pandas-stubs 2.0.2.230605
pandas 2.0.2
Additional context
Maybe FloatFormatType can be used (again) but could be updated to be more specific about the callable (Callable[[float], str]).
And perhaps the pandas docs should be updated? Unless support for passing format string was supposed to have been removed?
@twoertwein can you take a look at this?
And perhaps the pandas docs should be updated? Unless support for passing format string was supposed to have been removed?
In the past, it apparently did not work. Since it works now, it probably is intended to pass formatting strings. Do you mind opening an issue at pandas to clarify the intend/update the docs?
In the past, it apparently did not work.
Issue that seems related: https://github.com/pandas-dev/pandas/issues/9448
It still doesn't work for extension dtypes.
pd.Series([1.1, 2.2, 3.3]).to_string(float_format="%.3e") # works
pd.Series([1.1, 2.2, 3.3]).convert_dtypes().to_string(float_format="%.3e") # `TypeError: 'str' object is not callable`
Thank you, @zmoon ! In that case, I would be inclined to not add str to the annotations.
| gharchive/issue | 2023-06-13T19:39:34 | 2025-04-01T06:45:20.148944 | {
"authors": [
"Dr-Irv",
"twoertwein",
"zmoon"
],
"repo": "pandas-dev/pandas-stubs",
"url": "https://github.com/pandas-dev/pandas-stubs/issues/730",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1977050368 | core.groupby.generic.DataFrameGroupBy missing keyword-only arguments engine and engine_kwargs
Describe the bug
core.groupby.generic.DataFrameGroupBy missing keyword-only arguments
To Reproduce
Provide a minimal runnable pandas example that is not properly checked by the stubs.
Indicate which type checker you are using (mypy or pyright).
Show the error message received from that type checker while checking your example.
Current:
DataFrameGroupBy().aggregate("size", "some", "args", engine=0, engine_kwargs="not valid", other_kwarg="")
# no error despite `engine` as `engine_kwargs` not being valid types for those kw-only args
Expected:
mypy
error: No overload variant of "aggregate" of "DataFrameGroupBy" matches argument types "str", "str", "str", "int", "int", "str" [call-overload]
note: Possible overload variants:
note: def aggregate(self, arg: Literal['size']) -> Series[Any]
note: def aggregate(self, arg: Union[Union[Callable[..., Any], str, ufunc], list[Union[Callable[..., Any], str, ufunc]], Mapping[Any, Union[Union[Callable[..., Any], str, ufunc], list[Union[Callable[...,
Any], str, ufunc]]]]] = ..., *args: Any, engine: Optional[str] = ..., engine_kwargs: Optional[dict[str, bool]] = ..., **kwargs: Any) -> DataFrame
pyright
error: Argument of type "Literal[0]" cannot be assigned to parameter "engine" of type "str | None" in function "aggregate"
Type "Literal[0]" cannot be assigned to type "str | None"
"Literal[0]" is incompatible with "str"
Type cannot be assigned to type "None" (reportGeneralTypeIssues)
error: Argument of type "Literal[0]" cannot be assigned to parameter "engine_kwargs" of type "dict[str, bool] | None" in function "aggregate"
Type "Literal[0]" cannot be assigned to type "dict[str, bool] | None"
"Literal[0]" is incompatible with "dict[str, bool]"
Type cannot be assigned to type "None" (reportGeneralTypeIssues)
Please complete the following information:
OS: Windows
OS Version 10.0.19045 Build 19045
python version: 3.9.13
version of type checker:
mypy 1.6.1 (compiled: yes)
pyright@1.1.334
version of installed pandas-stubs: 2.1.1.230928
Additional context
Noticed while drilling down seaborn types to complete some method signatures in typeshed
I'm not sure that the code above is correct in terms of how pandas documents DataFrameGroupBy.aggregate(), because of the first 3 arguments, which would need to be a list to do what I think you want.
Having said that, the engine and engine_kwargs arguments are missing from the stubs.
For us to address this, we would need a more complete example.
| gharchive/issue | 2023-11-03T23:34:58 | 2025-04-01T06:45:20.156527 | {
"authors": [
"Avasam",
"Dr-Irv"
],
"repo": "pandas-dev/pandas-stubs",
"url": "https://github.com/pandas-dev/pandas-stubs/issues/810",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1167645919 | A server that also acts as a client
Please fill out the following system information before opening an issue:
OS (e.g. Ubuntu 18.04): MacOS Monterey 12.2.1
Go version (e.g. Go 1.13): go1.17.8 darwin/arm64
gnet version (e.g. v1.0.0): v2.0.1
What is your question about gnet?
Is there a way in gnet to start a UPD server and send a message to another server from the same FD without the server sending a message before.
An example: There is UDP server A and UDP server B. A and B have never communicated with each other before. A now wants to send a message to B.
How would I do something like this in gnet? Is it possible at all?
It is important that the port of UDP server A does not change when sending the message. So it is not possible to just create a new client, which then sends the message to B.
Couldn't you just create a UDP socket to server B in server A and bind that socket with the port you want?
Couldn't you just create a UDP socket to server B in server A and bind that socket with the port you want?
Maybe, but how? gnet.NewClient() and gnet.Client.Dial() does not allow to bind to the socket of the UDP server A.
It is not an option to create a new socket.
Is there a way to send network messages without using sockets?
Besides, I don't think you are allowed to reuse an in-use port of an existing server as the source port of sender.
u want it working like this ?
ServerC (port m) -- msg --> ServerA (port n) -- relay --> ServerB port(port p) -- rsp --> ServerA (port n)
u need to figure out a way send pkts to different peer from one local Conn object.
Does one udp Conn support customized peer addr on each asyncWrite ?
@KevinRoebert
I understand what you want. In fact, by using the builtin net package in Go to directly bind a UDP socket, one can read messages that are being listened to, or send them to another target. That might have allowed you to do what you want. However, the reason we use gnet is probably because it's more convenient than implementing directly with the builtin net package.
In gnet, internally, there is a file descriptor for a UDP socket bound to the server's port address in the event loop. In fact, it would be great if the gnet maintainer provided a convenient interface for this.
The workaround I found is as follows.
import (
"net"
"os"
"github.com/panjf2000/gnet"
)
type server struct {
*gnet.EventServer
conn *net.PacketConn
}
func (s *server) OnInitComplete(svr gnet.Server) (action gnet.Action) {
fd, err := svr.DupFd()
if err != nil {
log.Fatal(err)
}
file := os.NewFile(uintptr(fd), "udp")
conn, err := net.FilePacketConn(file)
if err != nil {
log.Fatal(err)
}
// net.UDPConn, actually
s.conn = &conn
log.Printf("Listening %s://%s", svr.Addr.Network(), svr.Addr)
return
}
...
And then it can be used elsewhere.
...
func (s *server) SendTo(p []byte, addr net.Addr) error {
if _, err := (*d.conn).WriteTo(p, addr); err != nil {
return err
}
return nil
}
In fact, because I am in a bit of a hurry, I had to make it work somehow. For the time being, I will use it this way, but as I mentioned before, I hope that gnet can make it more elegantly.
Thank you.
| gharchive/issue | 2022-03-13T17:40:31 | 2025-04-01T06:45:20.437878 | {
"authors": [
"HustCoderHu",
"KevinRoebert",
"civilizeddev",
"panjf2000"
],
"repo": "panjf2000/gnet",
"url": "https://github.com/panjf2000/gnet/issues/345",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
199421824 | Unclear when report is output
gulp-tslint version: 4.05
tslint version: 4.0.5
Operating system: Windows 10
Example gulp configuration (if applicable):
var files = gulp.src("src/**/*.ts", { base: "src" });
gutil.log("Begin TSLint Report");
files.pipe(tslint({
formatter: "verbose",
}))
.pipe(tslint.report({
emitError: false,
}));
gutil.log("End TSLint Report");
Error console output:
[13:28:35] Begin TSLint Report
[13:28:35] End TSLint Report
(no-empty-interface) common/lobby/IServerToClient.ts[1, 11]: An empty interface is equivalent to `{}`.
I would expect the output to be written during the pipe(tslint.report())
Your console logging is synchronous, that's why they're logged before the .report call.
| gharchive/issue | 2017-01-08T13:27:16 | 2025-04-01T06:45:20.553118 | {
"authors": [
"CoenraadS",
"panuhorsmalahti"
],
"repo": "panuhorsmalahti/gulp-tslint",
"url": "https://github.com/panuhorsmalahti/gulp-tslint/issues/100",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
225364654 | [WIP] Fixing build number
Fixing #88
:white_check_mark: Build ServiceBusExplorer 4.0.78 completed (commit https://github.com/paolosalvatori/ServiceBusExplorer/commit/ffa393301a by @SeanFeldman)
:white_check_mark: Build ServiceBusExplorer 0.0.82 completed (commit https://github.com/paolosalvatori/ServiceBusExplorer/commit/e5686556ad by @SeanFeldman)
:white_check_mark: Build ServiceBusExplorer build-86 completed (commit https://github.com/paolosalvatori/ServiceBusExplorer/commit/33efe76917 by @SeanFeldman)
:white_check_mark: Build ServiceBusExplorer build-88 completed (commit https://github.com/paolosalvatori/ServiceBusExplorer/commit/082f79e7e6 by @SeanFeldman)
| gharchive/pull-request | 2017-05-01T00:07:34 | 2025-04-01T06:45:20.559181 | {
"authors": [
"SeanFeldman"
],
"repo": "paolosalvatori/ServiceBusExplorer",
"url": "https://github.com/paolosalvatori/ServiceBusExplorer/pull/101",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2428542095 | engine2 enters SYNCING loop when handling old blocks
Running the following reth-bench command:
./target/release/reth-bench new-payload-fcu --rpc-url https://ethereum-holesky-rpc.publicnode.com --from 1986222 --to 1986224 --jwtsecret ~/Library/Application\ Support/reth/holesky/jwt.hex
On an engine2 node with the following latest_block:
2024-07-24T21:56:02.989520Z INFO reth::cli: Status connected_peers=0 freelist=5574 latest_block=1986617
I then get repeats for the following:
2024-07-24T20:51:19.690570Z DEBUG reth::cli: Event: Handler(ForkchoiceUpdated(ForkchoiceState { head_block_hash: 0x07f5afd31991775d6deebc23a913098634cf79a18a660fb5e4d6fd016fbfe152, safe_block_hash: 0xdea1af403b9ac006dd29c929006225dd365bf6d652bac4546923398bd2df4e2a, finalized_block_hash: 0x5fbc320f4a62c589fcd8441bb3074078dd5dba4aab563d1e4cc232931c88a807 }, Syncing))
2024-07-24T20:51:19.690660Z DEBUG reth::cli: Event: Handler(ForkchoiceUpdated(ForkchoiceState { head_block_hash: 0x07f5afd31991775d6deebc23a913098634cf79a18a660fb5e4d6fd016fbfe152, safe_block_hash: 0xdea1af403b9ac006dd29c929006225dd365bf6d652bac4546923398bd2df4e2a, finalized_block_hash: 0x5fbc320f4a62c589fcd8441bb3074078dd5dba4aab563d1e4cc232931c88a807 }, Syncing))
2024-07-24T20:51:19.690750Z DEBUG reth::cli: Event: Handler(ForkchoiceUpdated(ForkchoiceState { head_block_hash: 0x07f5afd31991775d6deebc23a913098634cf79a18a660fb5e4d6fd016fbfe152, safe_block_hash: 0xdea1af403b9ac006dd29c929006225dd365bf6d652bac4546923398bd2df4e2a, finalized_block_hash: 0x5fbc320f4a62c589fcd8441bb3074078dd5dba4aab563d1e4cc232931c88a807 }, Syncing))
And the benchmarker does not continue
this is fixed
| gharchive/issue | 2024-07-24T21:58:24 | 2025-04-01T06:45:20.591580 | {
"authors": [
"Rjected"
],
"repo": "paradigmxyz/reth",
"url": "https://github.com/paradigmxyz/reth/issues/9786",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1681632483 | feat: add missing pending block conversions
add a few missing conversions for pending block
Codecov Report
Merging #2372 (e60dbb7) into main (20e842b) will decrease coverage by 25.31%.
The diff coverage is 0.00%.
:mega: This organization is not using Codecov’s GitHub App Integration. We recommend you install it so Codecov can continue to function properly for your repositories. Learn more
@@ Coverage Diff @@
## main #2372 +/- ##
===========================================
- Coverage 71.73% 46.43% -25.31%
===========================================
Files 481 481
Lines 58420 58442 +22
===========================================
- Hits 41909 27138 -14771
- Misses 16511 31304 +14793
Flag
Coverage Δ
integration-tests
18.76% <0.00%> (-0.01%)
:arrow_down:
unit-tests
39.53% <0.00%> (-26.98%)
:arrow_down:
Flags with carried forward coverage won't be shown. Click here to find out more.
Impacted Files
Coverage Δ
crates/storage/provider/src/providers/mod.rs
0.00% <0.00%> (ø)
... and 213 files with indirect coverage changes
Help us with your feedback. Take ten seconds to tell us how you rate us. Have a feature suggestion? Share it here.
| gharchive/pull-request | 2023-04-24T16:18:46 | 2025-04-01T06:45:20.600358 | {
"authors": [
"codecov-commenter",
"mattsse"
],
"repo": "paradigmxyz/reth",
"url": "https://github.com/paradigmxyz/reth/pull/2372",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
82637628 | Implement Memory Map Feature
[This is a placeholder so we don't forget to implement this feature. The details will be placed here before we attempt solving this issue.]
F8 or Ctrl-M - Tokenize and open Memory Map display.
There is a good starting point with the legacy system to visualize the data. It would be interesting to audit the existing and see if modern convention affords us opportunity to refresh the view.
Yes, especially because we are also wanting this to support the Propeller in the future. We will have to make a mock up and list functionality that we can discuss.
| gharchive/issue | 2015-05-30T00:06:58 | 2025-04-01T06:45:20.604705 | {
"authors": [
"PropGit",
"urbantumbleweed"
],
"repo": "parallaxinc/ChromeIDE",
"url": "https://github.com/parallaxinc/ChromeIDE/issues/113",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2362788206 | The P.B Button wrapper for shad/cn component is not passing down props
Describe the bug
In my testing of DropdownMenu of Shad/cn from platform-bible-react lib, I noticed the PB Button is not passing down the props of Shad/cn.
To Reproduce
Use Button component from platform-bible-react as a child in the DropdownMenu.
It is not opening the menu as it is supposed to.
I used regular input component as a child and it is opening the menu.
<DropdownMenu>
<DropdownMenuTrigger asChild>
{/* // regular input is opening the menu */}
<input type="button" value="Click Me" />
{/* P.B Button is not opening the menu */}
{/* <Button>Shad DropdownMenu trigger</Button> */}
</DropdownMenuTrigger>
Expected behavior
Button component should open the menu.
The button is working as expected after the wrapper is removed. Verifies the fix worked by @jolierabideau .
Working as expected.
| gharchive/issue | 2024-06-19T16:47:37 | 2025-04-01T06:45:20.618503 | {
"authors": [
"roopa0222"
],
"repo": "paranext/paranext-core",
"url": "https://github.com/paranext/paranext-core/issues/948",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1477400375 | 🛑 api.parspec.io is down
In 0389196, api.parspec.io (https://api.parspec.io/admin/) was down:
HTTP code: 0
Response time: 0 ms
Resolved: api.parspec.io is back up in 10d1952.
| gharchive/issue | 2022-12-05T20:40:42 | 2025-04-01T06:45:20.634223 | {
"authors": [
"parikshit-parspec"
],
"repo": "parikshit-parspec/upptime",
"url": "https://github.com/parikshit-parspec/upptime/issues/71",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
650548554 | DOC Clarify release documentation
Reference Issue
Clarify release steps
What does this implement/fix? Explain your changes.
Any other comments?
cc @agramfort @maikia
thanks @lucyleeow
| gharchive/pull-request | 2020-07-03T11:53:27 | 2025-04-01T06:45:20.636031 | {
"authors": [
"lucyleeow",
"maikia"
],
"repo": "paris-saclay-cds/ramp-board",
"url": "https://github.com/paris-saclay-cds/ramp-board/pull/436",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
1595411529 | enable server running offline
#622 in conjunction with specifying a data URL for shiki should solve this
Fixed by #842
| gharchive/issue | 2023-02-22T16:27:12 | 2025-04-01T06:45:20.636874 | {
"authors": [
"harrysolovay",
"tjjfvi"
],
"repo": "paritytech/capi",
"url": "https://github.com/paritytech/capi/issues/640",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
852309600 | changed config format to allow multiple nodes for a single para config
Resolves https://github.com/paritytech/polkadot-launch/issues/74
allows elegantly expressing multiple nodes on the same parachain
some remaining linting fixes
replaced with https://github.com/paritytech/polkadot-launch/pull/87
| gharchive/pull-request | 2021-04-07T11:27:47 | 2025-04-01T06:45:20.667250 | {
"authors": [
"joelamouche",
"shawntabrizi"
],
"repo": "paritytech/polkadot-launch",
"url": "https://github.com/paritytech/polkadot-launch/pull/79",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1426760998 | [pallet_collective] Usage of Vec instead of BoundedVec
I found an instance, where other pallets are using a BoundedVec, but pallet_collective does not.
https://github.com/paritytech/substrate/blob/6195ef41946017ec860d2d8ad58f9bc18b5797bf/frame/collective/src/lib.rs#L221
I will like to take this issue up and fix. @Chralt98
The pallet bounding is ongoing here https://github.com/paritytech/substrate/issues/8629
In this case it is in the genesis config and therefore not a problem anyway.
| gharchive/issue | 2022-10-28T06:35:22 | 2025-04-01T06:45:20.675904 | {
"authors": [
"Chralt98",
"dharjeezy",
"ggwpez"
],
"repo": "paritytech/substrate",
"url": "https://github.com/paritytech/substrate/issues/12574",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1478874258 | Scheduler cleanup
Scheduler should remove any trailing None items in agendas as items get removed.
A migration should do this for preexisting items.
@gavofyork i will like to pick this up, So, are you saying the Agenda storage map should not store BoundedVec<Option<ScheduledOf<T>>, T::MaxScheduledPerBlock> anymore but store this instead
BoundedVec<ScheduledOf<T>, T::MaxScheduledPerBlock>?
Done in https://github.com/paritytech/substrate/pull/12989
This issue has been mentioned on Polkadot Forum. There might be relevant details there:
https://forum.polkadot.network/t/polkadot-release-analysis-v0-9-37/1736/1
| gharchive/issue | 2022-12-06T11:28:01 | 2025-04-01T06:45:20.678589 | {
"authors": [
"Polkadot-Forum",
"dharjeezy",
"gavofyork",
"ggwpez"
],
"repo": "paritytech/substrate",
"url": "https://github.com/paritytech/substrate/issues/12851",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1366617498 | Update ss58-registry
Update ss58 registry to include the latest additions to the chains list.
You can just do cargo update -p ss58-registry in your chain.
You can just do cargo update -p ss58-registry in your chain.
Thx for the tip :D
Mind getting this in too?
bot merge
| gharchive/pull-request | 2022-09-08T16:03:58 | 2025-04-01T06:45:20.680280 | {
"authors": [
"MOZGIII",
"bkchr",
"ggwpez"
],
"repo": "paritytech/substrate",
"url": "https://github.com/paritytech/substrate/pull/12220",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1642610102 | Remove HeaderBackend requirement from AuthorityDiscovery and NetworkWorker
NetworkWorker and AuthorityDiscovery took in a client that needed to implement HeaderBackend just to fetch the best hash/genesis hash. Shuffling around the code a bit can eliminate this client and allow some nice refactorings in cumulus.
cumulus companion: https://github.com/paritytech/cumulus/pull/2385
bot merge
| gharchive/pull-request | 2023-03-27T18:33:24 | 2025-04-01T06:45:20.681699 | {
"authors": [
"skunert"
],
"repo": "paritytech/substrate",
"url": "https://github.com/paritytech/substrate/pull/13730",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
507280509 | Better Parameterisation for Fee system
Makes is slightly more substrate-style by using Get instead of floating constant or hardcoded values.
[ ] Needs a Polkadot PR.
code is OK but wonder about API again maybe we would prefer not having this code in runtime/impl.rs as probably quite some other runtime will implement the same.
In contracts it is organized a bit differently and maybe we should inspiration from it:
on constracts::trait there is this
type ComputeDispatchFee: ComputeDispatchFee<<Self as Trait>::Call, BalanceOf<Self>>;
and then in contracts crate there is :
/// The default dispatch fee computor computes the fee in the same way that
/// the implementation of `ChargeTransactionPayment` for the Balances module does. Note that this only takes a fixed
/// fee based on size. Unlike the balances module, weight-fee is applied.
pub struct DefaultDispatchFeeComputor<T: Trait>(PhantomData<T>);
impl<T: Trait> ComputeDispatchFee<<T as Trait>::Call, BalanceOf<T>> for DefaultDispatchFeeComputor<T> {
fn compute_dispatch_fee(call: &<T as Trait>::Call) -> BalanceOf<T> {
...
}
}
So the dilemma is to leave it in impl.rs or move it to srml/transaction-payment.
In practice, I am fine with both. We are providing a default behaviour for something which at the end of the day the user should define.
But what I don't like about putting this targetedAdjustment in a substrate module is that it is too polkadot-specific. In otherwords, naming a polkadot requirement as the default of substrate is not good. Although, I have made the struct configurable here by being able to change the 25% value. But still the whole logic of "Changing fees based on congestion" seems polkadot-specific to me.
I am even tempted to wipe this ting from substrate node and only leave it in polkadot repo. Substrate node can just use () or ConvertInto, meaning that fees will not change over time.
But what I don't like about putting this targetedAdjustment in a substrate module is that it is too polkadot-specific. In otherwords, naming a polkadot requirement as the default of substrate is not good.
Agreed, I am pro putting this logic in Polkadot and using () for the Substrate node. I appreciate the separation and think the default for a Substrate node should be no change in fees over time. In this case, documentation could guide users to configure fees in a similar way to Polkadot as an example (@JoshOrndorff should be kept aware bc of https://github.com/substrate-developer-hub/recipes/pull/62), but I really don't like having the Polkadot slow-adjusting fee model as the default in Substrate...
I can expand on this, but I basically don't expect most developers building with Substrate to adopt the same fee structure as a relay chain. The economics of parachains/sovereign chains should and will be very different. Personally, I think something like BRAQ (section 4.2, page 9) or some similar model for subsidizing actions within a defined period of blocks might be more appropriate than charging per method call.
@thiolliere we are writing a doc about weights and I thought I would keep this Iced to apply very small changes to docs and prevent opening new PRs every day.
| gharchive/pull-request | 2019-10-15T14:24:17 | 2025-04-01T06:45:20.688920 | {
"authors": [
"4meta5",
"kianenigma",
"thiolliere"
],
"repo": "paritytech/substrate",
"url": "https://github.com/paritytech/substrate/pull/3823",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
545668585 | client: allow reverting finalized blocks
This PR extends the revert functionality to allow reverting past finalized blocks. It also fixes the handling of leaves when reverting blocks.
In which case it'll need to revert finalized blocks?
@kigawas https://medium.com/@gavofyork/kusamas-first-adventure-2cd4f439a7a4 A writeup of why we needed this for Kusama.
| gharchive/pull-request | 2020-01-06T11:00:34 | 2025-04-01T06:45:20.690836 | {
"authors": [
"andresilva",
"kigawas"
],
"repo": "paritytech/substrate",
"url": "https://github.com/paritytech/substrate/pull/4535",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
935939604 | Bump linregress due to security vulnerability
https://rustsec.org/advisories/RUSTSEC-2021-0070
Here is the cargo deny failure:
= Advisory: https://rustsec.org/advisories/RUSTSEC-2021-0070
= The `Deserialize` implementation for `VecStorage` did not maintain the invariant that the number of elements must equal `nrows * ncols`. Deserialization of specially crafted inputs could allow memory access beyond allocation of the vector.
This flaw was introduced in v0.11.0 ([`086e6e`](https://github.com/dimforge/nalgebra/commit/086e6e719f53fecba6dadad2e953a487976387f5)) due to the addition of an automatically derived implementation of `Deserialize` for `MatrixVec`. `MatrixVec` was later renamed to `VecStorage` in v0.16.13 ([`0f66403`](https://github.com/dimforge/nalgebra/commit/0f66403cbbe9eeac15cedd8a906c0d6a3d8841f2)) and continued to use the automatically derived implementation of `Deserialize`.
This flaw was corrected in commit [`5bff536`](https://github.com/dimforge/nalgebra/commit/5bff5368bf38ddfa31416e4ae9897b163031a513) by returning an error during deserialization if the number of elements does not exactly match the expected size.
= Announcement: https://github.com/dimforge/nalgebra/issues/883
= Solution: Upgrade to >=0.27.1
= nalgebra v0.26.2
└── statrs v0.14.0
└── linregress v0.4.2
└── frame-benchmarking v3.1.0
└── pallet-balances v3.0.0
I don't think I can add a label here, so can someone label this as "insubstantial" since it is only a dependency bump?
| gharchive/pull-request | 2021-07-02T17:12:34 | 2025-04-01T06:45:20.692655 | {
"authors": [
"trevor-crypto"
],
"repo": "paritytech/substrate",
"url": "https://github.com/paritytech/substrate/pull/9262",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
143720375 | Oh wow you have to change the name.
I suggest canibekikked. As I understand it, "kike" is to jews what "nigger" is to black people.
That is a ridiculous level of attempted political correctness, and works counter to preventing words like kike or nigger from having negative connotations.
If anyone wants to claim being offended by this totally unrelated usage of letters arranged in the same order then they are simply ensuring they can be offended in the future.
CanIBeKiked / canibekiked is fine.
According to the definition in Urban Dictionary, I think "kiked" is appropriate: https://www.urbandictionary.com/define.php?term=Kiked
"Appropriate" as in "wow so inappropriate", you mean?
Also, the correct past tense of "kik" is "kikked". Because it's pronounced "kick", not "kike".
Please, look at my avatar. Of course I think it's appropriate.
Oh, I couldn't make out the letters before. Or know what Trump stands for, but I'm guessing from the context that he's big on racism.
Is 'please look at my avatar' the new 'please clap' ?
That is a ridiculous level of attempted political correctness, and works counter to preventing words like kike or nigger from having negative connotations.
Insensitivity to the fact that those words do have negative connotations, and will for many more years does not advance anything positive.
If anyone wants to claim being offended by this totally unrelated usage of letters arranged in the same order then they are simply ensuring they can be offended in the future.
That sounds like you're suggesting that the problem lies entirely with those offended - you're effectively victim-blaming.
This isn't just pointless political correctness. It's basic sensitivity to the fact that words mean things.
There are no victims here. The application does not do anything negative for anybody, the naming is coincidental and if anyone was to attempt to be offended the barest bit of context would suffice to stifle it.
There should be a github reaction called "oy vey another shoah".
Well, I am UNSUBBING from this thread of butthurt individuals. Sorry guys.
@egroat I kinda understand your viewpoint as well, but what gets me is that "kiked" is the past tense of "kike", not of "kik".
@skorokithakis ok I can see that. I still think its an overreaction but couched as it is with the semi-justification as a correction, there is reason for a renaming.
If you consider the fact that "can I be jewed" would be an anti-semtic way to say "can I be cheated," it's possible for a reasonable person to think "can I be kiked" is an intentional double entendre with this name. I agree this really needs to be changed.
A few years ago my Hungarian grandmother died. In Hungarian there's a word for "wooden" which sounds like "fuh". And a lot of words starting with a " k" sound.
One of my favorite childhood pasttimes was coming up with words like "wooden duck" (fuh kah cha) or wooden car (fuh kar ay) or wooden spoon (fuh ka nal).
Oh how I laughed and laughed at how unrelated things sounded hilarious in English. But the moral of this story is that Hungarians should change their language to protect those who only hear the imagined English that was never really there.
Kik is pronounced like "kick". "Kiked" does not contain the pronounciation of "kick", but does contain the pronounciation of "kike", which is an ethnic slur. Probably a mistake by someone who isn't a native English speaker, but it could be taken as an attempted double entendre.
Also regarding the comment above mine, "can I be" is English, not Hungarian. The name/phrase is intended to be read like an English sentence.
Also, the author is in for a great surprise when he sees this comment thread. @parro-it, just to put you at ease, none of us really feel very strongly about this, it's up to you.
I feel strongly about it.
Why not just canibekikd?
Personally, I'm not offended by the name because I don't think it was intentional -- however, it's unnecessarily ambiguous and controversial. I know this isn't very academic, but take a look at how the words kiked, kikked, and kik'd are used on Twitter.
My vote is for "kikked", which removes the confusion of potential anti-semitic without adding an apostrophe (unless you want this to be parro-it/canibekik-apostrophe-d). Either way, good project -- thanks for your work on this @parro-it.
Add functionality to see if your repo name has a racial slur?
Hi all. I'm so sorry, I had no idea what "kike" means before reading this issue.
I'm not a native speaker, I intend the word only in the sense "to suffer the same fate as the azer/kik project".
I will rename the repository and add a note in readme.
Add functionality to see if your repo name has a racial slur?
I saw some day ago a repo containg a list of offensive words.
I think it should be easy to check if a repo name is contained there.
@parro-it No problem, I just figured I'd give you a heads up because not many people know what "kike" means unless they're native English speakers or watch South Park.
@skorokithakis, thank you for the issue! I used to watch South Park, but it's translated in Italian here.
@parro-it FWIW I've never heard the term until now and I'm a native English speaker. Might be regional? If you find that repo with a list of words to avoid, could you post it here? It'd help me out, maybe others :)
Sure, it's https://github.com/shutterstock/List-of-Dirty-Naughty-Obscene-and-Otherwise-Bad-Words
brb going to open an issue with kik interactive claiming kik sounds like "kike".
You're all grasping for straws here. There's nothing wrong with the name. Don't make a mountain out of a molehill. Actually not even a molehill. This is the friggen plains, flat as shit. No mountains. Stop.
@joshmanders Just to be clear, you're advocating that we revert the name to a misspelling that also happens to be a racial slur? Hilarious. I understand your frustration against people who want to be offended by everything, but unfortunately this is nothing more than a non-native English speaker making an innocent misspelling.
Don't believe me? Try replacing "kik" with similar verbs:
Beg: "beged" or "begged"?
Clap: "claped" or "clapped"?
Fan: "faned" or "fanned"?
Whip: "whiped" or "whipped"?
Now, let's try "kik" again:
Kik: "kiked" or "kikked"?
If you're looking for a primer on the English past tense (maybe you're learning English for the first time?), I'd highly recommend this video. There's a great example at the 2-minute mark where they go over the past tense of "stop" (which is "stopped", not "stoped").
Please let me know if there are any other questions I can answer for you regarding English, the past tense, or just general etiquette on GitHub.
@christianbundy Then we all better be outraged that GitHub, Inc, or GitLab, or any other company, including Torvalds the creator of Git change the names right now, because Git is offensive. http://www.urbandictionary.com/define.php?term=Git
Can't pick and choose what you want to be offended about. Either be full blown politically correct or stop taking so much offense to every little possible thing.
The difference there is that someone can't just assume the word "git" is directed at them, because "git" means "an unpleasant or contemptible person". "Kike" is offensive to any jew. I'm as much against political correctness as the next guy, but, since a repo rename is just one setting away, I don't see what's so bad about this. It's like someone naming their repo "canibeniggered" (completely by accident) and people being outraged at other people's soft suggestion that it perhaps be renamed.
I didn't strong-arm anyone into it or even express any strong preference, and @parro-it was gracious enough to rename it when he became aware of the issue, so I don't think this is the particular hill to fight the anti-PC battle on...
30+ comments of people being outraged by something that shouldn't have any outrage around it. When applied to the context in which this module was made, there is no reason one should even think about "kike". It's those who read into it wrong who are blowing it up.
How about we do more productive things. The JavaScript ecosystem is too entangled with being politically correct about every single little aspect of it, that I don't even know if anyone actually does anything outside of scour code to find something to be offended and outraged about.
@joshmanders I've previously expressed that I'm not offended by the name, but at least now there's no doubt as to whether you read the discussion. My vote to change the name came from the fact that it's incorrectly spelled, just as I would do if the name said "stoped" instead of "stopped".
The only reason you want to revert to the incorrect spelling is that you're offended that other people are offended by "kike", and want to make a point of keeping the racial slur to stick it to "politically correct" people.
If node-gyp took over gyp and a project named can-i-be-gypped project changed its name to avoid using "gypped", then your comments might be relevant. Unfortunately for you, that isn't what happened. Instead, there was a misspelling in the project name, and it happened to also be a racial slur.
Why are you still commenting? You've already demonstrated:
You didn't read the previous discussion.
You aren't offended by "kike".
If anyone is offended by something you aren't, they're "politically correct".
You're offended by "politically correct people". (?!?!)
The things that offend you are the only things that are offensive, everything else is just "making a mountain out of friggen plains, flat as shit".
You want to purposefully butcher the English language just so that you can offend the "politically correct" people with racial slurs.
I don't know which holy war you think you're fighting, but you aren't going to win it by misspelling words to create racial slurs (however if that were a game, I'd call it "Donald Trump Scrabble").
If you're going to use the misspelling as your basis for changing it, then just say that. But alas, most of the comments here are about it being a racial slur.
I'm not fighting any holy war, I'm just expressing my deep disgust with this ecosystems vast desire to remove anything that could offend anyone for any little thing, all while using something that is widely known as being an offensive slur. But that don't matter apparently. Hypocrisy is the war I am fighting.
30+ comments of people being outraged by something that shouldn't have any outrage around it.
Nobody is actually outraged. We're merely capable of seeing things from perspectives other than our own.
How about we do more productive things. The JavaScript ecosystem is too entangled with being politically correct about every single little aspect of it, that I don't even know if anyone actually does anything outside of scour code to find something to be offended and outraged about.
And you're meta-outraged at this, being whatever the opposite of politically correct is, instead of actually doing anything outside of scouring comments to find something to be offended about.
And for what it's worth, gyp is also questionable, but so are a billion things. Just because someone is choosing to comment on X, but not on Y, doesn't make their comments on X irrelevant or hypocritical.
This isn't just pointless political correctness. It's basic sensitivity to the fact that words mean things.
Let me preface this with while I think the name of this module was fine how it was, I don't advocate calling anyone an N word, or a kike, but you're 100% right on what you said words mean things. They meant many things long before their negative forms became something. Just like the swastika meant peace before Nazi germany. The N word is derived from how it was pronounced/said in Latin, and Kike came from the Yiddish word "circle" for which is how Jewish immigrants signed on Ellis Island.
It's when used in a certain context that it goes from being used in it's correct and original form, to being used in a negative form.
It's when used in a certain context that it goes from being used in it's correct and original form, to being used in a negative form.
Unfortunately, history weighs words down with meaning even when they're used in a neutral context. Come on - you didn't even write "nigger", choosing instead "N word". From context, it would be 100% obvious that you wouldn't be using it to be offensive, but the word is still, in 2016, weighed down so heavily with connotation that it's hard to type, to see, to read.
From context, it would be 100% obvious that you wouldn't be using it to be offensive, but the word is still, in 2016, weighed down so heavily with connotation that it's hard to type, to see, to read.
Exactly, "nigger" is used way more in society than "kike" is, that's why I have a hard time typing it and resort to the "N word". Looks I'm not going against you guys on the term being offensive. I'm going against the fact that in such a context, people are making a scruff out of it BEING offensive. @christianbundy I apologize to you because after going back and rereading again (read this thread on and off over the day or so because of events away from computer keeping me) I see you were fighting for the correct spelling.
It bothers me heavily how much people are finding anything they can to be offended about. And it's always the ones who have no reason to be offended who are offended. Why? Why are people that way? Then they're only offended about what's higher profile offensive. As if someone somewhere is holding them hostage and forcing them to be offended or some skeleton is going to be let out of the closet and that can't happen.
Why are white people being offended by someone saying the N word? But every black person I've talked to seriously about this (including many close personal friends) say it doesn't bother them?
Why are non-Jewish people being offended by a non-native English speaker misspelling a word and the resulting sequence when pronounced incorrectly is a slur?
Why is it okay for all these overly offended but shouldn't have a reason to be offended people get to express their offense to such things, yet someone who's sick of seeing the constant "OMG THAT'S OFFENSIVE TO SOMEONE SOMEWHERE PROBABLY SO YOU SHOULD SPEND 600 HOURS CAREFULLY CRAFTING YOUR 3 WORD SENTENCE INCASE SOMEONE IS OFFENDED BY LONG WORDS!" but I can't express my offense to people being offended so much?
I'm grumpy, I apologize to everyone. Have a good Sunday evening.
@joshmanders You're good, man. I understand where you're coming from -- even if I disagree on this specific issue. Thought you might like this.
FWIW "git" refers to Torvalds himself:
Torvalds quipped about the name git...: "I'm an egotistical bastard, and I name all my projects after myself. First 'Linux', now 'git'."
https://en.wikipedia.org/wiki/Git_(software)#History
kikked? You got cucked, cuck.
@skorokithakis what about https://github.com/skorokithakis/expounder/ ? Ex pounder? I think we've seen lighter names being targeted.
I wish this sort of issues was reported by people that were actually offended rather than those thinking someone else might.
@rr- Please just let this issue die. Go write a blog post or something productive, but don't waste our time harassing people on a closed issue that has nothing to do with you. Comments like yours are the reason that issues get locked.
I need coffee.
I wish this sort of issues was reported by people that were actually offended rather than those thinking someone else might
Isn't this the equivalent of saying that bugs should only be reported in production once they affect someone?
@rr- Expounder was intentional, that lib just likes to pound all your exes.
@pavellishin Eh, it's more like saying "this isn't a bug unless someone minds the behaviour", which I kind of agree with. In any case, this has become a much larger issue than I intended, so can we just drop it, like @christianbundy said?
Plus, none of you cottoned on to my real agenda, which is that I'm a prescriptivist and hated the malformed past tense.
What if you just change it to "canibejewed"?
| gharchive/issue | 2016-03-26T15:42:32 | 2025-04-01T06:45:20.741253 | {
"authors": [
"Gargron",
"Incognito",
"OBAMATRON",
"PixnBits",
"athomann",
"christianbundy",
"egroat",
"gyardley",
"johnstt",
"joshmanders",
"parro-it",
"pavellishin",
"rr-",
"ryanbreen",
"shotxxxx",
"skorokithakis",
"tilleps",
"wodim"
],
"repo": "parro-it/canibekikked",
"url": "https://github.com/parro-it/canibekikked/issues/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
206516694 | Implement custom equality for Source comparison
Resolves #130: Correct cycle detection for Sub tokens depends on Source equality. Implementing equals() and hashCode() correctly for Source means also implementing it for all Tokens, ValueExpressions, Expressions, Encoding, ParseItems and all concrete Sources. This PR completes this operation.
Resolves #148: As part of this, the interface to ValueExpression has changed to no longer require an Environment as input to its eval() method, but just the ParseGraph. This to make sure objects not related to the parse state are not inadvertently used to determine equality (e.g., Callbacks).
Resolves #147: A bug was also fixed where Sub used the Source of its address ValueExpression result as input instead of the input from the Environment.
Additionally, CurrentOffset has been removed, as it depended on a field of the Environment being passed (offset). It has been replaced by a oneliner in Shorthand that calculates it based on the ParseGraph. The new implementation is not strictly the same but should be equivalent in most cases.
ConstantSlice has also been removed. Its inner class ConstantSource has been promoted to a top-level class and the regular Slice implementation can be used to create a slice of a ConstantSource.
Testing of both equals() and hashCode() is done using two new test classes. AutoEqualityTest is an automated, combinatorial approach for most classes. Custom testing of some specific cases is in EqualityTest, along with a specific cycle detection case for Sub.
I still cannot see the Better Code Hub results, what is it complaining about?
This code needs to be refactored
Massive improvements! Thanks for the great feedback @rdvdijk!
Approved, pending the Travis build :computer:
| gharchive/pull-request | 2017-02-09T14:17:20 | 2025-04-01T06:45:20.774283 | {
"authors": [
"jvdb",
"rdvdijk"
],
"repo": "parsingdata/metal",
"url": "https://github.com/parsingdata/metal/pull/149",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
1806017232 | Cannot resolve partykit/server when using TypeScript ESM (moduleResolution: NodeNext)
My partykit version: 0.0.0-6927442
My tsconfig:
{
"compilerOptions": {
"strict": true,
"noUncheckedIndexedAccess": true,
"target": "ES2020",
"moduleResolution": "NodeNext",
"module": "NodeNext"
}
}
My code:
import { PartyKitServer, PartyKitRoom, PartyKitContext } from "partykit/server";
Result:
Could not find a declaration file for module 'partykit/server'. '/Users/tom.sherman/code/partykit-xstate-demo/node_modules/.pnpm/partykit@0.0.0-6927442/node_modules/partykit/dist/server.js' implicitly has an 'any' type.
Try `npm i --save-dev @types/partykit` if it exists or add a new declaration (.d.ts) file containing `declare module 'partykit/server';`
Thanks for catching this!! We need to fix this, open to PRs. Possibly need to fix our package.json, but I'm not sure just now how to do so.
ok, I think this should be fixed, please reinstall partykit@beta and try again
Thanks, ESM is cursed!
| gharchive/issue | 2023-07-15T10:54:34 | 2025-04-01T06:45:20.804543 | {
"authors": [
"threepointone",
"tom-sherman"
],
"repo": "partykit/partykit",
"url": "https://github.com/partykit/partykit/issues/167",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1707897363 | Error: NetworkError when attempting to fetch resource.
Keep getting this error every time I upload.
All endpoints seem to be correct.
Got it, file size too large!
| gharchive/issue | 2023-05-12T16:21:35 | 2025-04-01T06:45:20.805762 | {
"authors": [
"alayhi"
],
"repo": "pashpashpash/vault-ai",
"url": "https://github.com/pashpashpash/vault-ai/issues/75",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
367821791 | Deprecated FB fields in strategy.js cause Facebook error when called up as profileFields
Nice work, thank you for the fork from passport-facebook!
There is one issue I found related to the field names Facebook has changed over the time.
There is the following code in lib/strategy.js which attempts to fill in some user variables:
Strategy.prototype._convertProfileFields = function(profileFields) {
var map = {
'id': 'id',
'username': 'username',
'displayName': 'name',
'name': ['last_name', 'first_name', 'middle_name'],
'gender': 'gender',
'profileUrl': 'link',
'emails': 'email',
'photos': 'picture'
};
When I use profileFields in the FacebookStrategy, like this:
passport.use(new FacebookStrategy({
clientID: config.facebook_api_key,
clientSecret:config.facebook_api_secret ,
callbackURL: config.callback_url,
graphApiVersion: 'v3.1',
profileFields: ['id', 'displayName', 'emails', 'profileUrl']
},
the outdated code gets called, and Facebook comes back with the message below:
FacebookGraphAPIError: (#12) username field is deprecated for versions v2.0 and higher
at /home/ubuntu/expresslogin/node_modules/@passport-next/passport-facebook/lib/strategy.js:168:21
at passBackControl (/home/ubuntu/expresslogin/node_modules/oauth/lib/oauth2.js:132:9)
at IncomingMessage. (/home/ubuntu/expresslogin/node_modules/oauth/lib/oauth2.js:157:7)
at emitNone (events.js:111:20)
at IncomingMessage.emit (events.js:208:7)
at endReadableNT (_stream_readable.js:1064:12)
at _combinedTickCallback (internal/process/next_tick.js:138:11)
at process._tickCallback (internal/process/next_tick.js:180:9)
Just letting you know. If the fields are not called up, everything is ok, like this:
passport.use(new FacebookStrategy({
clientID: config.facebook_api_key,
clientSecret:config.facebook_api_secret ,
callbackURL: config.callback_url,
graphApiVersion: 'v3.1'
},
And the app can grab these fields directly from Facebook using another library.
:+1: I'll take a look I thought I'd sorted all these apparently not!
| gharchive/issue | 2018-10-08T14:38:48 | 2025-04-01T06:45:20.812632 | {
"authors": [
"bendaklara",
"rwky"
],
"repo": "passport-next/passport-facebook",
"url": "https://github.com/passport-next/passport-facebook/issues/2",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
246429523 | React Bootstrap Tooltips?
Is there a good way to hook up tooltips if we're currently using React-Bootstrap?
eg:
https://react-bootstrap.github.io/components.html#tooltips
Semantic UI is also a good pick to go
@VictorChen add this prop to your CalendarHeatmap component.
transformDayElement={(rect, value, index) => {
const tooltip = (
<Tooltip id="tooltip">{value ? value.date : ''}</Tooltip>
);
return (
<OverlayTrigger placement="top" overlay={tooltip}>
{rect}
</OverlayTrigger>
);
}}
@scotthelps Nice!! did not know about transformDayElement prop, thanks!
@cassiopagnoncelli could you post a snippet of how you got the tooltip to work with semantic-ui-react popup component?
@grahambryan Honestly I couldn't, it was a mere suggestion. However it seems to be as easy as wrapping Semantic UI's around
https://github.com/patientslikeme/react-calendar-heatmap/blob/0fb7a61638addea1e1d9254b55a777748a716c9d/src/index.jsx#L251
@cassiopagnoncelli Thanks! I ended up going with react-tooltip
<CalendarHeatmap endDate={new Date('2018-01-01')} startDate={this.state.startDate} values={this.state.values} onClick={this.onClick} classForValue={(value) => { if (!value) { return 'color-empty'; } returncolor-github-${value.count}; }} tooltipDataAttrs={customTooltipDataAttrs} /> <ReactTooltip type='warning' effect='solid' />
const customTooltipDataAttrs = (value) => ({ 'data-tip': value.date === null ? '' : value.status + ' on ' + value.date });
and using a controlled semantic-UI-react modal that will open onClick and give the user more detailed information about the day clicked.
| gharchive/issue | 2017-07-28T18:48:33 | 2025-04-01T06:45:20.831289 | {
"authors": [
"VictorChen",
"cassiopagnoncelli",
"grahambryan",
"scotthelps"
],
"repo": "patientslikeme/react-calendar-heatmap",
"url": "https://github.com/patientslikeme/react-calendar-heatmap/issues/32",
"license": "mit",
"license_type": "permissive",
"license_source": "bigquery"
} |
2098948320 | Enrichments
Redo enrichments across all flags.
Consider the application of the ML approach to predict likely activity in different assay systems in lieu of the boolean approach to assign activity.
Addressed.
| gharchive/issue | 2024-01-24T19:35:20 | 2025-04-01T06:45:20.832463 | {
"authors": [
"patlewig"
],
"repo": "patlewig/nts_pfas",
"url": "https://github.com/patlewig/nts_pfas/issues/1",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
266464496 | Limit the height of the selected options
Good Morning,
I am trying to limit the height of the column of selected elements, in order to make it scrollable, I have come to use in the inspector even the element.style selector, forcing the height even with important.
I have not been able to make that listing fit the desired height.
Could you help me?
Thx so much!
Luisfermp
Ok, there was an extra display attribute that shouldn't have been there. That's fixed now, all you need to do is add a height attribute to the .selected and selections CSS rule.
| gharchive/issue | 2017-10-18T11:51:19 | 2025-04-01T06:45:20.837870 | {
"authors": [
"Luisfermp18",
"patosai"
],
"repo": "patosai/tree-multiselect.js",
"url": "https://github.com/patosai/tree-multiselect.js/issues/49",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
531925069 | Their is a way to add list of "filters" and "remove"(X) button?
The idea/behavior:
https://www.carmax.com/cars/all
By mixitup-multifilter
Hi Ezra,
By default, MixItUp includes UI for simple buttons only. If you want to do anything more complex, you can do so - but you need to create the UI, logic, and handle events yourself - which would control MixItUp via its API.
Thanks.
| gharchive/issue | 2019-12-03T11:41:48 | 2025-04-01T06:45:20.843429 | {
"authors": [
"Ezra-Siton-UIX",
"patrickkunka"
],
"repo": "patrickkunka/mixitup",
"url": "https://github.com/patrickkunka/mixitup/issues/530",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
305031165 | building error
Could not get unknown property 'bintray_user' for object of type com.jfrog.bintray.gradle.BintrayExtension. Open File
Why are you cloning the project when you can use the gradle installation? Are planning to contribute? The bintray_user is my bintray user name and bintray_key is my password. They are stored local on my computer and not supposed to be exported with the project.
| gharchive/issue | 2018-03-14T05:48:54 | 2025-04-01T06:45:20.849553 | {
"authors": [
"daydream11",
"patriques82"
],
"repo": "patriques82/alphavantage4j",
"url": "https://github.com/patriques82/alphavantage4j/issues/23",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
659607866 | Cpfed 3735 megamenu tray
name: pfe-nav mega menu tray and single column dropdown
labels: "feature"
pfe-navigation (CPFED 3735 megamenu tray)
Added mega menu tray styles and pfe-cta examples in mega menu tray. Added single column dropdown class and example of single col dropdown menu.
@wesruv you can remove this line from the index.html "../../pfe-dropdown/dist/pfe-dropdown.umd.js" I just had that there to review the pfe-dropdown on the page.
Merging
Please squash when merging and ensure your commit message uses conventional commit formatting.
This is still early days and we're working in a branch for the epic, I'm merging this into our working branch so I can work on it with @kelsS' work.
| gharchive/pull-request | 2020-07-17T20:24:26 | 2025-04-01T06:45:20.858732 | {
"authors": [
"kelsS",
"wesruv"
],
"repo": "patternfly/patternfly-elements",
"url": "https://github.com/patternfly/patternfly-elements/pull/985",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
426654302 | refactor(flex-layout): convert device breakpoints to grid breakpoints
fix #1654
Deploy preview for pf-next ready!
Built with commit d9d13bdca44b7acc52ea7ad5d0dc0df79f11b7fe
https://deploy-preview-1657--pf-next.netlify.com
looks good! could you also remove the breakpoints from scss-variables.scss?
:tada: This PR is included in version 2.0.0 :tada:
The release is available on:
npm package (@latest dist-tag)
GitHub release
Your semantic-release bot :package::rocket:
| gharchive/pull-request | 2019-03-28T18:49:39 | 2025-04-01T06:45:20.862528 | {
"authors": [
"mattnolting",
"mcoker",
"patternfly-build"
],
"repo": "patternfly/patternfly-next",
"url": "https://github.com/patternfly/patternfly-next/pull/1657",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2736981058 | Instance backup missing some data?
I recently did a fresh install to try and get around the python 3.12 issue with octolapse and arcwelder, but after restoring my previous backup to the new instance it's missing all of my uploaded gcode files, timelapses, and within octoprint backups. Is this normal behavior for the octoprint_deploy instance backups? If I had known I would have saved the within octoprint backup to a usb or something, I lost a lot of data to this.
I tried making a new instance and using the same backup, but this time it included the Gcode files, but not timelapses or achievements. not really sure what data is saved in the backup, and what isn't.
It excludes timelapses, but everything else should be there.
| gharchive/issue | 2024-12-12T22:15:49 | 2025-04-01T06:45:20.900179 | {
"authors": [
"Big-Channie",
"paukstelis"
],
"repo": "paukstelis/octoprint_deploy",
"url": "https://github.com/paukstelis/octoprint_deploy/issues/183",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
2561196945 | Uncaught ImportError in http_handler.py line 7
Aider version: 0.58.1
Python version: 3.12.6
Platform: Linux-6.11.1-2-cachyos-x86_64-with-glibc2.40
Python implementation: CPython
Virtual environment: Yes
OS: Linux 6.11.1-2-cachyos (64bit)
Git version: git version 2.46.2
An uncaught exception occurred:
Traceback (most recent call last):
File "aider", line 8, in <module>
sys.exit(main())
^^^^^^
File "main.py", line 537, in main
main_model = models.Model(
^^^^^^^^^^^^^
File "models.py", line 646, in __init__
res = self.validate_environment()
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "models.py", line 845, in validate_environment
res = litellm.validate_environment(model)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "llm.py", line 23, in __getattr__
self._load_litellm()
File "llm.py", line 30, in _load_litellm
self._lazy_module = importlib.import_module("litellm")
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "__init__.py", line 90, in import_module
return _bootstrap._gcd_import(name[level:], package, level)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "<frozen importlib._bootstrap>", line 1387, in _gcd_import
File "<frozen importlib._bootstrap>", line 1360, in _find_and_load
File "<frozen importlib._bootstrap>", line 1331, in _find_and_load_unlocked
File "<frozen importlib._bootstrap>", line 935, in _load_unlocked
File "<frozen importlib._bootstrap_external>", line 995, in exec_module
File "<frozen importlib._bootstrap>", line 488, in _call_with_frames_removed
File "__init__.py", line 8, in <module>
from litellm.llms.custom_httpx.http_handler import AsyncHTTPHandler, HTTPHandler
File "http_handler.py", line 7, in <module>
from httpx import USE_CLIENT_DEFAULT
ImportError: cannot import name 'USE_CLIENT_DEFAULT' from 'httpx' (/home/dmtarmey/MyProjects/aider-env/lib/python3.12/site-packages/httpx/__init__.py)
Thank you for filing this issue.
Can you please try to install or reinstall aider in a separate Python environment? Either with venv or pipx?
Please remove the existing environment/aider installation first, e.g. with pipx:
$ pipx uninstall aider-chat
uninstalled aider-chat! ✨ 🌟 ✨
$ pipx install aider-chat
installed package aider-chat 0.58.1, installed using Python 3.12.6
These apps are now globally available
- aider
done! ✨ 🌟 ✨
This document may be helpful -> https://aider.chat/docs/troubleshooting/imports.html
| gharchive/issue | 2024-10-02T10:18:07 | 2025-04-01T06:45:20.939806 | {
"authors": [
"DMTarmey",
"fry69"
],
"repo": "paul-gauthier/aider",
"url": "https://github.com/paul-gauthier/aider/issues/1882",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
306319331 | AdvertisementData contains empty values on iOS
Hi,
On iOS, I'm not seeing any data for the following properties on AdvertisementData in my ScanResult event stream handler:
localName
manufacturerData
serviceData
I observed this to be consistent over ScanResults for responses for 19 different devices.
localName returns an empty String, and manufacturerData and serviceData return an empty List<int>.
Let me know if I can provide more details or if I can help in troubleshooting, debugging, or verifying a fix.
Thanks.
Confirming this behavior. The information is key, especially serviceData, otherwise you have to connect to devices to determine what services they support.
Hey Paul,
I am now able to confirm that there's advertising packet data that I can't get to from flutter_blue.
My BLE device is include a Service UUID in its advertising packet, which will allow my app to know that a given service is available BEFORE connecting. I'm guessing this should be coming through in ScanResult.advertisementData.serviceData ...
I can see this data coming from my device using various apps such as Nordic's nRF Connect, where it looks like this:
When I access this and various other properties of AdvertisementData for my device in flutter_blue, all are empty. When I look at the same info for "all devices in range" I see just one device out of ~20 that shows a non-empty value for serviceData. I am able to see localName for my device. (I don't see other values but those aren't being sent).
I can provide more info as needed.
Paul, following up on this one --
hi,
i'm trying to create an android and ios app in flutter for ble operations, i need to take all bytes of advertisement data while scanning time and i need to make a condition check based on the advertisement bytes.
i can get all byte using NRF CONNECT (advertisement packet 1 and 2 total is 62 byte data),
how to get from this?
could you please help me,..
This error is also happening on Android. I'm using the example app to run on API 22.
I just downloaded the example and ran it.
@jack4git Could you share the "raw" table, like shown by @francisraj ?
Hi,
Now we can get rssi and manufacturer data, but we need full bytes of
advertisement data ( it is maybe 61 bytes data).
And how to make broadcastUpdate like android coding.
@Override
public void onCharacteristicRead(BluetoothGatt gatt,
BluetoothGattCharacteristic characteristic,
int status) {
if (status == BluetoothGatt.GATT_SUCCESS) {
* broadcastUpdate(ACTION_DATA_AVAILABLE, characteristic);
//>>>>>>>>>>>> Like this.*
}
}
using this broadcastupdate i can easily find and filter to take correct
data from ble.
On Sat, Aug 11, 2018 at 1:48 PM, Paul DeMarco notifications@github.com
wrote:
@jack4git https://github.com/jack4git Sorry for the delay. If you're
still in need of a fix, please try out the latest version 0.4.0 on pub and
let me know how it goes.
Thanks!
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
https://github.com/pauldemarco/flutter_blue/issues/54#issuecomment-412260044,
or mute the thread
https://github.com/notifications/unsubscribe-auth/Amzt_5rePJLV9BUGIzxuN2Lvu_gh2cH0ks5uPpNugaJpZM4SviGO
.
--
Regards,
Francis.ARM | Software Engineer.
@ryanhanks Just confirming, is this issue fixed for you?
Paul,
I am also facing same issue . After scan Service data is null . ( flutter_blue , Master -
Latest commit 6f6fa30 )
Thanks,
Gaurav
Hi, I have the same issue. using the latest build. Any updates?
Hi
I have found that the ScanResult serviceData is overwritten by the devices scan response service data if it contains values.
Ideally ScanResult AdvertisingData would return both the ScanResult and ScanResponse service data in separate Lists. I guess this depends on how the OS's handle the Scan Response data though.
I hope this helps someone
Pete
Hi,
having the same issue but on Android. ServiceData and manufacturerData are empty for one of 3 different devices.
Using Pixel 2 with Android 10 and Flutter 1.17.1
having the same issue but on Android 4.4.4, nRF Connect shows this data
Anyone find anything out on thiis?
any fix guys? this is an example:
Android (works fine) :
iOS:
same here :
| gharchive/issue | 2018-03-19T03:07:38 | 2025-04-01T06:45:20.962739 | {
"authors": [
"AymenXD",
"GauravPatni",
"PeteSwad",
"Yinten",
"francisraj",
"hbsis-eduardo",
"jack4git",
"novas1r1",
"pauldemarco",
"recvec",
"robomike",
"ryanhanks"
],
"repo": "pauldemarco/flutter_blue",
"url": "https://github.com/pauldemarco/flutter_blue/issues/54",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
579147062 | Merge pull request #1 from pauljamescleary/master
Updating copy
Updating copy
Hi @yeleywale, thanks for trying out the repository. If you're trying to update your local fork you would do so using git on your computer.
This article might be helpful to you.
Have a great day.
| gharchive/pull-request | 2020-03-11T10:11:03 | 2025-04-01T06:45:20.979500 | {
"authors": [
"yeleywale",
"zakpatterson"
],
"repo": "pauljamescleary/scala-pet-store",
"url": "https://github.com/pauljamescleary/scala-pet-store/pull/341",
"license": "apache-2.0",
"license_type": "permissive",
"license_source": "bigquery"
} |
48122441 | es6-shim breaks Chrome App CSP
Unless sandboxed, es6-shim cannot run inside of Chrome Apps right now due to this line, which triggers the CSP, and is functionally an eval() statement.
I'm unsure if the best fix is reverting https://github.com/paulmillr/es6-shim/commit/ceeb51cc99bbe3729e96eae8c7af966a2d67dfde, or if a better solution can be found.
CSP is quite handy and — in addition to being used in Chrome apps and Firefox OS — many web applications also use it. We're one of them and just ran into this.
Would anthonyryan1's method work? [https://github.com/paulmillr/es6-shim/issues/301#issuecomment-64966395]
@sandstrom See my response right below it: https://github.com/paulmillr/es6-shim/issues/301#issuecomment-64973036
I think I've got a possible fix here - https://gist.github.com/stewart/321560330b152dfde649.
Didn't cause any issues in the environments I tested in, but maybe I'm missing something?
@stewart What did you test it on? It would need to work back to IE 6, in node 0.6 and later, all io.js versions, node-webkit/nw.js, web workers, the node vm module's sandbox, and probably in all of a CJS, AMD, and no-module environment for each.
The one major drawback to my approach that the CSP-violating version doesn't have is that if es6-shim gets wrapped up in an IIFE w/ strict mode enabled (say by a minifier or project bundler), it'll stop working.
On the other hand, a situation like that is easier to work around for developers than a CSP-violating dependency, and relatively easy for the library to detect (check if globals === undefined).
Travis indicates at least Node/iojs support about on par with master - https://travis-ci.org/stewart/es6-shim/builds/69499282, I'll see about testing other platforms later (don't have access to IEs older than 10 though, I'm afraid). I did test in Chrome/Safari/Firefox and they all seemed to like it.
I'm less concerned about browsers and node (where window and global are reliable enough), than I am about web workers and node-webkit, both of which we can't currently test on travis, nor do I have any way to test locally.
I don't have test harnesses available for either at present, but I'll see about getting something worked together over the weekend, unless someone else beats me to it.
Thanks!
I've worked on this with @rwaldron and @jugglinmike in person, and we've found a non-eval approach that seems to work in browsers, workers, node-webkit, and node.
In order to unblock Chrome apps from having CSP issues, I'm going to go ahead with this change. Please let me know if this change has broken anything.
@ljharb this throws an error when running using WSH.
Altough, es6-shim does not work in WSH anyway...
Thanks, that's good to know - what is the thrown error?
If you can provide me with various exceptions that it throws (as separate issues, please) I'll do my best to fix them!
throw new Error('unable to locate global object');
Perfect, that's exactly what it's supposed to do :-) It appears that in WSH, the only option is to get this in the global context. http://stackoverflow.com/questions/14450424/equivalent-to-window-in-jscript-runtime
I'd suggest then that in wsh, users need to do something like var global = this; in the global context prior to including the es6-shim.
@ljharb , setTimeout is not supported, so
var promiseSupportsSubclassing = supportsSubclassing(globals.Promise, function(S) {
return S.resolve(42).then(function() {}) instanceof S;
});
promiseSupportsSubclassing test throws an error...
@Yaffle Try the latest commit in WSH - please file a new issue for me if there's any further issues.
@ljharb , well... it works with var global = this; and es5-shim, but throws some errors:
https://github.com/paulmillr/es6-shim/blob/master/es6-shim.js#L1155
"Number" is a bad name for IE8 (probablye because of http://kiro.me/blog/nfe_dilemma.html), so before this line is reached, Number is called and throws the error, because isBinary is not defined yet.
wrapConstructor uses Object.getOwnPropertyNames, which is not supported in IE8 event with es5-shim;
Map and Set are not polyfilled in IE8 because of if (supportsDescriptors) {
@Yaffle Can you please file an issue for 1 and 2? 3 is expected and desired.
@ljharb Using var global = this; before loading es6-shim works with Nashorn too. Thanks!
| gharchive/issue | 2014-11-07T19:44:26 | 2025-04-01T06:45:20.996683 | {
"authors": [
"Yaffle",
"chmx",
"ljharb",
"sandstrom",
"stewart"
],
"repo": "paulmillr/es6-shim",
"url": "https://github.com/paulmillr/es6-shim/issues/301",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
476336376 | For statement doesn't loop?
https://github.com/paulmillr/es6-shim/blob/master/es6-shim.js#L1055
Is this a mistake? Surely just an if?
Yes, looks to be a think-o dating from f77f9f52ccc4765a0674a001fd1f258fdd7cf58a. The current spec is https://www.ecma-international.org/ecma-262/6.0/#sec-%arrayiteratorprototype%.next and doesn't have a for loop. (I think the for loop was from some intermediate stage where we were expecting to skip entries in sparse arrays.)
Of course it loops - it just doesn't have an initializer, because var i = this.i is already done higher up in the function.
There certainly may be a better way to express it, but there's nothing broken about it.
Ok, I don't think that I'd describe it as a loop as it returns unconditionally, I wasn't suggesting that it was broken more a strange use of for.
Going to close this; but if there's a compelling reason to refactor that code (including "is much clearer to read") then a PR would be welcome!
| gharchive/issue | 2019-08-02T20:04:42 | 2025-04-01T06:45:21.000282 | {
"authors": [
"Xotic750",
"cscott",
"ljharb"
],
"repo": "paulmillr/es6-shim",
"url": "https://github.com/paulmillr/es6-shim/issues/458",
"license": "MIT",
"license_type": "permissive",
"license_source": "github-api"
} |
1402400292 | primeiro commit
criei um ariquivo e adicionei frase
Não gostei
| gharchive/pull-request | 2022-10-09T21:16:25 | 2025-04-01T06:45:21.001141 | {
"authors": [
"k3vinrich4rd",
"pauloren7"
],
"repo": "pauloren7/TreinoBranchsTop",
"url": "https://github.com/pauloren7/TreinoBranchsTop/pull/1",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
1915249833 | Added toggle for right sidebar
Hello!
In this MR I added the toggle to collapse and expand the right sidebar because it takes up a lot of space.
When working on a small laptop screen, large tasks are hard to read
Valery, please fix the conflicts!
| gharchive/pull-request | 2023-09-27T10:46:59 | 2025-04-01T06:45:21.018244 | {
"authors": [
"KozyarValeriy",
"pavelpower"
],
"repo": "pavelpower/jira-helper",
"url": "https://github.com/pavelpower/jira-helper/pull/8",
"license": "Apache-2.0",
"license_type": "permissive",
"license_source": "github-api"
} |
2433883139 | Can not add conda executable in pycharm-ce
Description
Can not add conda executable in pycharm-ce.
Already did everything described in README.
Versions
❯ pixi --version
pixi 0.26.1
❯ pixi list | grep pixi-pycharm
pixi-pycharm 0.0.6 unix_1234567_0 8.6 KiB conda pixi-pycharm-0.0.6-unix_1234567_0.conda
PyCharm version: PyCharm 2024.1.4 (Community Edition)
PC-241.18034.82,2024年6月25日 构建
17.0.11+1-b1207.24 amd64
VM: OpenJDK 64-Bit Server VM,JetBrains s.r.o.
Platform: Debian 12
Linux zjc-office 6.1.0-23-amd64 #1 SMP PREEMPT_DYNAMIC Debian 6.1.99-1 (2024-07-15) x86_64 GNU/Linux
Logs
Attempting to run: ['info', '--envs', '--json']
Traceback (most recent call last):
File "/home/zjc/python_works/seiser_pixi/.pixi/envs/default/libexec/conda", line 165, in <module>
sys.exit(main(args))
^^^^^^^^^^
File "/home/zjc/python_works/seiser_pixi/.pixi/envs/default/libexec/conda", line 139, in main
conda_info_envs_json()
File "/home/zjc/python_works/seiser_pixi/.pixi/envs/default/libexec/conda", line 79, in conda_info_envs_json
"envs": pixi_envs(),
^^^^^^^^^^^
File "/home/zjc/python_works/seiser_pixi/.pixi/envs/default/libexec/conda", line 57, in pixi_envs
platform = pixi_json(["info", *std_pixi_args()])["platform"]
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/zjc/python_works/seiser_pixi/.pixi/envs/default/libexec/conda", line 53, in pixi_json
return json.loads(pixi([*cmd, "--json"]))
^^^^^^^^^^^^^^^^^^^^^^
File "/home/zjc/python_works/seiser_pixi/.pixi/envs/default/libexec/conda", line 22, in pixi
out = subprocess.check_output(["pixi", *cmd])
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/zjc/python_works/seiser_pixi/.pixi/envs/default/lib/python3.11/subprocess.py", line 466, in check_output
return run(*popenargs, stdout=PIPE, timeout=timeout, check=True,
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/zjc/python_works/seiser_pixi/.pixi/envs/default/lib/python3.11/subprocess.py", line 548, in run
with Popen(*popenargs, **kwargs) as process:
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/zjc/python_works/seiser_pixi/.pixi/envs/default/lib/python3.11/subprocess.py", line 1026, in __init__
self._execute_child(args, executable, preexec_fn, close_fds,
File "/home/zjc/python_works/seiser_pixi/.pixi/envs/default/lib/python3.11/subprocess.py", line 1955, in _execute_child
raise child_exception_type(errno_num, err_msg, err_filename)
FileNotFoundError: [Errno 2] No such file or directory: 'pixi'
https://github.com/pavelzw/pixi-pycharm/issues/16#issuecomment-2066359437 might fix this. Where is your pixi installed? If it's in ~/.pixi/bin/pixi, #20 is the issue which I should fix at some point.
#16 (comment) might fix this. Where is your pixi installed? If it's in ~/.pixi/bin/pixi, #20 is the issue which I should fix at some point.
Thank you very much! I found the solution works in #16 , I installed as default in ~/.pixi. Sorry to bother you, have a nice day.
| gharchive/issue | 2024-07-28T11:24:47 | 2025-04-01T06:45:21.024145 | {
"authors": [
"Zhou-Jiangcheng",
"pavelzw"
],
"repo": "pavelzw/pixi-pycharm",
"url": "https://github.com/pavelzw/pixi-pycharm/issues/29",
"license": "BSD-3-Clause",
"license_type": "permissive",
"license_source": "github-api"
} |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.