Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 2
665
| labels
stringlengths 4
554
| body
stringlengths 3
235k
| index
stringclasses 6
values | text_combine
stringlengths 96
235k
| label
stringclasses 2
values | text
stringlengths 96
196k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
682,273
| 23,339,136,811
|
IssuesEvent
|
2022-08-09 12:41:54
|
Polkadex-Substrate/Polkadex-Open-Beta
|
https://api.github.com/repos/Polkadex-Substrate/Polkadex-Open-Beta
|
closed
|
Site is currently vulnerable to click jacking
|
Unconfirmed FrontEnd immediate fix high priority Critical
|
**By raising this issue, I agree to the terms and conditions presented by Polkadex here:**[Terms and Conditions](https://github.com/Polkadex-Substrate/Polkadex-Open-Beta-Feedback/blob/master/Polkadex_Open_Beta_Program_-_Terms__Conditions.pdf)
**Description**
To check this, you can use this tool: http://online.attacker-site.com/html5/ClickjackingTester/
Typing in the URL for the orderbook or polkadex.trade site will confirm this. The easiest way to fix this would be to configure the appropriate X-frame response headers
onchainaddress:{esmTYAxXyJM4siS5nimmqeWcriYGEumnMc7Egofp8KGt8eJKd}
|
1.0
|
Site is currently vulnerable to click jacking - **By raising this issue, I agree to the terms and conditions presented by Polkadex here:**[Terms and Conditions](https://github.com/Polkadex-Substrate/Polkadex-Open-Beta-Feedback/blob/master/Polkadex_Open_Beta_Program_-_Terms__Conditions.pdf)
**Description**
To check this, you can use this tool: http://online.attacker-site.com/html5/ClickjackingTester/
Typing in the URL for the orderbook or polkadex.trade site will confirm this. The easiest way to fix this would be to configure the appropriate X-frame response headers
onchainaddress:{esmTYAxXyJM4siS5nimmqeWcriYGEumnMc7Egofp8KGt8eJKd}
|
non_infrastructure
|
site is currently vulnerable to click jacking by raising this issue i agree to the terms and conditions presented by polkadex here description to check this you can use this tool typing in the url for the orderbook or polkadex trade site will confirm this the easiest way to fix this would be to configure the appropriate x frame response headers onchainaddress
| 0
|
2,084
| 3,508,480,131
|
IssuesEvent
|
2016-01-08 18:02:50
|
elmsln/elmsln
|
https://api.github.com/repos/elmsln/elmsln
|
closed
|
Issue Importing Book Outline (XML)
|
bug courses / mooc infrastructure online / CIS
|
Hi. I'm getting the following warning messages when I try importing:

Is this something you've seen before?
|
1.0
|
Issue Importing Book Outline (XML) - Hi. I'm getting the following warning messages when I try importing:

Is this something you've seen before?
|
infrastructure
|
issue importing book outline xml hi i m getting the following warning messages when i try importing is this something you ve seen before
| 1
|
32,524
| 26,753,002,527
|
IssuesEvent
|
2023-01-30 21:15:37
|
microsoft/TypeScript
|
https://api.github.com/repos/microsoft/TypeScript
|
closed
|
tsserverlibrary.js includes reference to `require` that causes webpack to complain
|
Infrastructure Fix Available
|
# Bug Report
<!--
Please fill in each section completely. Thank you!
-->
### ๐ Search Terms
- webpack
- `require`
- `tsserverlibrary`
### ๐ Version & Regression Information
5.0.0-dev.20230130
### Problem
The latest `tsserverlibrary.js` contains a line:
```ts
var __require = /* @__PURE__ */ ((x) => typeof require !== "undefined" ? require : typeof Proxy !== "undefined" ? new Proxy(x, {
get: (a, b) => (typeof require !== "undefined" ? require : a)[b]
}) : x)(function(x) {
if (typeof require !== "undefined")
return require.apply(this, arguments);
throw new Error('Dynamic require of "' + x + '" is not supported');
});
```
This causes webpack to complain:
```
WARNING in ../node_modules/typescript/lib/tsserverlibrary.js 21:75-82
Critical dependency: require function is used in a way in which dependencies cannot be statically extracted
@ ./web/webServer.ts 32:24-65
```
As far as I can tell, `__require` isn't used anywhere in the file either. Can we remove it?
/cc @jakebailey
|
1.0
|
tsserverlibrary.js includes reference to `require` that causes webpack to complain - # Bug Report
<!--
Please fill in each section completely. Thank you!
-->
### ๐ Search Terms
- webpack
- `require`
- `tsserverlibrary`
### ๐ Version & Regression Information
5.0.0-dev.20230130
### Problem
The latest `tsserverlibrary.js` contains a line:
```ts
var __require = /* @__PURE__ */ ((x) => typeof require !== "undefined" ? require : typeof Proxy !== "undefined" ? new Proxy(x, {
get: (a, b) => (typeof require !== "undefined" ? require : a)[b]
}) : x)(function(x) {
if (typeof require !== "undefined")
return require.apply(this, arguments);
throw new Error('Dynamic require of "' + x + '" is not supported');
});
```
This causes webpack to complain:
```
WARNING in ../node_modules/typescript/lib/tsserverlibrary.js 21:75-82
Critical dependency: require function is used in a way in which dependencies cannot be statically extracted
@ ./web/webServer.ts 32:24-65
```
As far as I can tell, `__require` isn't used anywhere in the file either. Can we remove it?
/cc @jakebailey
|
infrastructure
|
tsserverlibrary js includes reference to require that causes webpack to complain bug report please fill in each section completely thank you ๐ search terms webpack require tsserverlibrary ๐ version regression information dev problem the latest tsserverlibrary js contains a line ts var require pure x typeof require undefined require typeof proxy undefined new proxy x get a b typeof require undefined require a x function x if typeof require undefined return require apply this arguments throw new error dynamic require of x is not supported this causes webpack to complain warning in node modules typescript lib tsserverlibrary js critical dependency require function is used in a way in which dependencies cannot be statically extracted web webserver ts as far as i can tell require isn t used anywhere in the file either can we remove it cc jakebailey
| 1
|
31,800
| 26,134,757,073
|
IssuesEvent
|
2022-12-29 10:35:52
|
firebase/flutterfire
|
https://api.github.com/repos/firebase/flutterfire
|
closed
|
Move all example apps to use "flutterfire-e2e-test" project
|
type: infrastructure
|
1. Move all example apps to use "flutterfire-e2e-tests" project.
2. Ensure platforms (Web, macOS, iOS & Android) run fine on the example app (i.e. no exceptions occur when building/running application as Podfile/build.gradle needs updating and the API works as intended).
Some example apps will need the Flutter implementation updating when moving to "flutterfire-e2e-tests" project such as the Dynamic Link. It is associated with the "react-native-firebase-testing" project.
## Plugins
Check the platform once it has been updated, it is running without issue & merged into "master" branch.
Firestore
- [ ] iOS
- [ ] Android
- [ ] macOS
- [ ] web
App Check
- [x] iOS
- [x] Android
- [x] macOS
- [x] web
Messaging
- [x] iOS
- [x] Android
- [x] macOS
- [x] web
Storage
- [x] iOS
- [x] Android
- [x] macOS
- [x] web
Remote Config
- [x] iOS
- [x] Android
- [x] macOS
- [x] web
Analytics
- [x] iOS
- [x] Android
- [x] macOS
- [x] web
Cloud Functions
- [x] iOS
- [x] Android
- [x] macOS
- [x] web
Auth
- [x] iOS
- [x] Android
- [x] macOS
- [x] web
App Installations
- [x] iOS
- [x] Android
- [x] macOS
- [x] web
Database
- [x] iOS
- [x] Android
- [x] macOS
- [x] web
Core
- [x] iOS
- [x] Android
- [x] macOS
- [x] web
Crashlytics
- [x] iOS
- [x] Android
- [x] macOS
Dynamic Links
- [x] iOS
- [x] Android
In App Messaging
- [X] iOS
- [X] Android
Performance
- [x] iOS
- [x] Android
- [x] web
Storage
- [x] iOS
- [x] Android
- [x] macOS
- [x] web
|
1.0
|
Move all example apps to use "flutterfire-e2e-test" project - 1. Move all example apps to use "flutterfire-e2e-tests" project.
2. Ensure platforms (Web, macOS, iOS & Android) run fine on the example app (i.e. no exceptions occur when building/running application as Podfile/build.gradle needs updating and the API works as intended).
Some example apps will need the Flutter implementation updating when moving to "flutterfire-e2e-tests" project such as the Dynamic Link. It is associated with the "react-native-firebase-testing" project.
## Plugins
Check the platform once it has been updated, it is running without issue & merged into "master" branch.
Firestore
- [ ] iOS
- [ ] Android
- [ ] macOS
- [ ] web
App Check
- [x] iOS
- [x] Android
- [x] macOS
- [x] web
Messaging
- [x] iOS
- [x] Android
- [x] macOS
- [x] web
Storage
- [x] iOS
- [x] Android
- [x] macOS
- [x] web
Remote Config
- [x] iOS
- [x] Android
- [x] macOS
- [x] web
Analytics
- [x] iOS
- [x] Android
- [x] macOS
- [x] web
Cloud Functions
- [x] iOS
- [x] Android
- [x] macOS
- [x] web
Auth
- [x] iOS
- [x] Android
- [x] macOS
- [x] web
App Installations
- [x] iOS
- [x] Android
- [x] macOS
- [x] web
Database
- [x] iOS
- [x] Android
- [x] macOS
- [x] web
Core
- [x] iOS
- [x] Android
- [x] macOS
- [x] web
Crashlytics
- [x] iOS
- [x] Android
- [x] macOS
Dynamic Links
- [x] iOS
- [x] Android
In App Messaging
- [X] iOS
- [X] Android
Performance
- [x] iOS
- [x] Android
- [x] web
Storage
- [x] iOS
- [x] Android
- [x] macOS
- [x] web
|
infrastructure
|
move all example apps to use flutterfire test project move all example apps to use flutterfire tests project ensure platforms web macos ios android run fine on the example app i e no exceptions occur when building running application as podfile build gradle needs updating and the api works as intended some example apps will need the flutter implementation updating when moving to flutterfire tests project such as the dynamic link it is associated with the react native firebase testing project plugins check the platform once it has been updated it is running without issue merged into master branch firestore ios android macos web app check ios android macos web messaging ios android macos web storage ios android macos web remote config ios android macos web analytics ios android macos web cloud functions ios android macos web auth ios android macos web app installations ios android macos web database ios android macos web core ios android macos web crashlytics ios android macos dynamic links ios android in app messaging ios android performance ios android web storage ios android macos web
| 1
|
14,048
| 4,807,807,376
|
IssuesEvent
|
2016-11-02 22:40:00
|
BowlerHatLLC/vscode-nextgenas
|
https://api.github.com/repos/BowlerHatLLC/vscode-nextgenas
|
closed
|
Not stopping at breakpoints in main class constructor
|
bug vscode-dependency
|
Ok, so I managed to get the compiling working on my mac, it now produces the bin folder and the output js files. The debugger executes the compile task then fires the chrome browser and the script runs but it doesn't stop at the breakpoints. I'm trying to debug using the local file option as per instructions.
|
1.0
|
Not stopping at breakpoints in main class constructor - Ok, so I managed to get the compiling working on my mac, it now produces the bin folder and the output js files. The debugger executes the compile task then fires the chrome browser and the script runs but it doesn't stop at the breakpoints. I'm trying to debug using the local file option as per instructions.
|
non_infrastructure
|
not stopping at breakpoints in main class constructor ok so i managed to get the compiling working on my mac it now produces the bin folder and the output js files the debugger executes the compile task then fires the chrome browser and the script runs but it doesn t stop at the breakpoints i m trying to debug using the local file option as per instructions
| 0
|
516,668
| 14,986,074,455
|
IssuesEvent
|
2021-01-28 20:46:15
|
xournalpp/xournalpp
|
https://api.github.com/repos/xournalpp/xournalpp
|
opened
|
GTK4 preparation
|
enhancement priority::low refactoring
|
GTK 4 was released last month and will probably become widely available over the next year or two. There's no rush to port to GTK 4 given the huge number of other important issues in our issue tracker, but it would be nice to [at least replace the deprecated GTK 3 parts that we can right now](https://developer.gnome.org/gtk4/4.0/gtk-migrating-3-to-4.html).
At least one low-hanging fruit I know of is the `gtk_pointer_grab` deprecation warning (or something along those lines) that's been around for over a year.
|
1.0
|
GTK4 preparation - GTK 4 was released last month and will probably become widely available over the next year or two. There's no rush to port to GTK 4 given the huge number of other important issues in our issue tracker, but it would be nice to [at least replace the deprecated GTK 3 parts that we can right now](https://developer.gnome.org/gtk4/4.0/gtk-migrating-3-to-4.html).
At least one low-hanging fruit I know of is the `gtk_pointer_grab` deprecation warning (or something along those lines) that's been around for over a year.
|
non_infrastructure
|
preparation gtk was released last month and will probably become widely available over the next year or two there s no rush to port to gtk given the huge number of other important issues in our issue tracker but it would be nice to at least one low hanging fruit i know of is the gtk pointer grab deprecation warning or something along those lines that s been around for over a year
| 0
|
28,961
| 23,631,056,153
|
IssuesEvent
|
2022-08-25 09:21:13
|
UnitTestBot/UTBotJava
|
https://api.github.com/repos/UnitTestBot/UTBotJava
|
opened
|
Disable container's monitoring metrics
|
infrastructure
|
**Description**
Decide whether to disable container's monitoring metrics or leave it as is.
**Expected behavior**
Not applicable
**Environment**
Not applicable
**Potential alternatives**
We can leave monitoring as is but I noted that we usually look on host's metrics and ignore container metrics. It's because container metrics don't have some unique data that host's metrics don't
**Context**
Not applicable
|
1.0
|
Disable container's monitoring metrics - **Description**
Decide whether to disable container's monitoring metrics or leave it as is.
**Expected behavior**
Not applicable
**Environment**
Not applicable
**Potential alternatives**
We can leave monitoring as is but I noted that we usually look on host's metrics and ignore container metrics. It's because container metrics don't have some unique data that host's metrics don't
**Context**
Not applicable
|
infrastructure
|
disable container s monitoring metrics description decide whether to disable container s monitoring metrics or leave it as is expected behavior not applicable environment not applicable potential alternatives we can leave monitoring as is but i noted that we usually look on host s metrics and ignore container metrics it s because container metrics don t have some unique data that host s metrics don t context not applicable
| 1
|
263,884
| 23,087,108,068
|
IssuesEvent
|
2022-07-26 12:25:27
|
elastic/beats
|
https://api.github.com/repos/elastic/beats
|
closed
|
[Filebeat] Flaky test_clean_inactive test
|
Filebeat flaky-test Team:Integrations Stalled
|
Seen in https://beats-ci.elastic.co/job/elastic+beats+pull-request+multijob-darwin/3672/beat=filebeat,label=macosx/testReport/junit/test_registrar/Test/test_clean_inactive/ (darwin build)
```
Traceback (most recent call last):
File "/System/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/unittest/case.py", line 331, in run
testMethod()
File "/private/var/lib/jenkins/workspace/elastic+beats+pull-request+multijob-darwin/beat/filebeat/label/macosx/src/github.com/elastic/beats/filebeat/tests/system/test_registrar.py", line 817, in test_clean_inactive
assert len(data) == 2
AssertionError
```
|
1.0
|
[Filebeat] Flaky test_clean_inactive test - Seen in https://beats-ci.elastic.co/job/elastic+beats+pull-request+multijob-darwin/3672/beat=filebeat,label=macosx/testReport/junit/test_registrar/Test/test_clean_inactive/ (darwin build)
```
Traceback (most recent call last):
File "/System/Library/Frameworks/Python.framework/Versions/2.7/lib/python2.7/unittest/case.py", line 331, in run
testMethod()
File "/private/var/lib/jenkins/workspace/elastic+beats+pull-request+multijob-darwin/beat/filebeat/label/macosx/src/github.com/elastic/beats/filebeat/tests/system/test_registrar.py", line 817, in test_clean_inactive
assert len(data) == 2
AssertionError
```
|
non_infrastructure
|
flaky test clean inactive test seen in darwin build traceback most recent call last file system library frameworks python framework versions lib unittest case py line in run testmethod file private var lib jenkins workspace elastic beats pull request multijob darwin beat filebeat label macosx src github com elastic beats filebeat tests system test registrar py line in test clean inactive assert len data assertionerror
| 0
|
579,105
| 17,172,819,761
|
IssuesEvent
|
2021-07-15 07:41:52
|
forkwhilefork/bird-config-generator
|
https://api.github.com/repos/forkwhilefork/bird-config-generator
|
closed
|
Add communities on filtered routes
|
High Priority enhancement
|
Add communities on filtered routes corresponding to the reasons they were filtered so that we can identify that reason later.
|
1.0
|
Add communities on filtered routes - Add communities on filtered routes corresponding to the reasons they were filtered so that we can identify that reason later.
|
non_infrastructure
|
add communities on filtered routes add communities on filtered routes corresponding to the reasons they were filtered so that we can identify that reason later
| 0
|
737,803
| 25,532,413,550
|
IssuesEvent
|
2022-11-29 09:27:38
|
mozilla/addons-server
|
https://api.github.com/repos/mozilla/addons-server
|
closed
|
Rename flagged by Scanners queue
|
component: reviewer tools priority: p3
|
The queue named "Flagged by Scanners" contains versions that have `needs_human_review=True` but we're using that flag for other-non-scanner related reasons now so the queue needs a rename.
|
1.0
|
Rename flagged by Scanners queue - The queue named "Flagged by Scanners" contains versions that have `needs_human_review=True` but we're using that flag for other-non-scanner related reasons now so the queue needs a rename.
|
non_infrastructure
|
rename flagged by scanners queue the queue named flagged by scanners contains versions that have needs human review true but we re using that flag for other non scanner related reasons now so the queue needs a rename
| 0
|
18,034
| 12,741,230,100
|
IssuesEvent
|
2020-06-26 05:25:37
|
acts-project/acts
|
https://api.github.com/repos/acts-project/acts
|
reopened
|
Add example project with Acts as external dependency
|
Impact - Minor Improvement Infrastructure Stale
|
[GitLab issue: #674](https://gitlab.cern.ch/acts/acts-core/issues/674)
We currently do not test that installed CMake configs work. We should have a small example project that builds a (non-trivial) example executable using Acts via the `find_package(...)` mechanism. This should look as simple as a single directory in the repository:
```
ExampleProject\
ExampleProject\CMakeLists.txt
ExampleProject\Propagate.cpp
```
Having this directly in the repository would allow us to build this in the CI using the installed build products from one of the Core build jobs. In principle, this could also be done as a separate project. Gitlab supports multi-project pipelines but only in the premium edition. This is currently not an option for us.
|
1.0
|
Add example project with Acts as external dependency - [GitLab issue: #674](https://gitlab.cern.ch/acts/acts-core/issues/674)
We currently do not test that installed CMake configs work. We should have a small example project that builds a (non-trivial) example executable using Acts via the `find_package(...)` mechanism. This should look as simple as a single directory in the repository:
```
ExampleProject\
ExampleProject\CMakeLists.txt
ExampleProject\Propagate.cpp
```
Having this directly in the repository would allow us to build this in the CI using the installed build products from one of the Core build jobs. In principle, this could also be done as a separate project. Gitlab supports multi-project pipelines but only in the premium edition. This is currently not an option for us.
|
infrastructure
|
add example project with acts as external dependency we currently do not test that installed cmake configs work we should have a small example project that builds a non trivial example executable using acts via the find package mechanism this should look as simple as a single directory in the repository exampleproject exampleproject cmakelists txt exampleproject propagate cpp having this directly in the repository would allow us to build this in the ci using the installed build products from one of the core build jobs in principle this could also be done as a separate project gitlab supports multi project pipelines but only in the premium edition this is currently not an option for us
| 1
|
12,080
| 18,882,484,364
|
IssuesEvent
|
2021-11-15 01:08:03
|
ossarioglu/SWE573-repo
|
https://api.github.com/repos/ossarioglu/SWE573-repo
|
opened
|
Create Mockup Drawings for Scenarios
|
requirements_analysis priority_medium size_medium
|
Create screenshots and drawings to visualize scenarios. Drawings should illustrate mentioned scenarios.
|
1.0
|
Create Mockup Drawings for Scenarios - Create screenshots and drawings to visualize scenarios. Drawings should illustrate mentioned scenarios.
|
non_infrastructure
|
create mockup drawings for scenarios create screenshots and drawings to visualize scenarios drawings should illustrate mentioned scenarios
| 0
|
560,977
| 16,608,023,061
|
IssuesEvent
|
2021-06-02 07:30:26
|
aviferdman/Workshop-on-Software-Engineering-Project
|
https://api.github.com/repos/aviferdman/Workshop-on-Software-Engineering-Project
|
closed
|
ืืืืืฉ ืงืืืื ื
|
features priority- critical
|
ืืืืืฉ ืงืืืื ื ืืคื ืืืจืืฉืืช ืืืืคืืขืืช ืืืจืกื 2:
ืืคืืื ืืืืืืฉ ืืืฉืง ืืฉืชืืฉ ื ืื
ืืืฉืง ืืฉืชืืฉ ืืืขืืื ืขืืืจ ืื ืกืื ืืฉืชืืฉ
ืืขืื.
ืืฉ ืืืืง ืืฉืืื ืื ืืชืชื ืืฉืืืืช ืืื ืืื ืืืืขื.
|
1.0
|
ืืืืืฉ ืงืืืื ื - ืืืืืฉ ืงืืืื ื ืืคื ืืืจืืฉืืช ืืืืคืืขืืช ืืืจืกื 2:
ืืคืืื ืืืืืืฉ ืืืฉืง ืืฉืชืืฉ ื ืื
ืืืฉืง ืืฉืชืืฉ ืืืขืืื ืขืืืจ ืื ืกืื ืืฉืชืืฉ
ืืขืื.
ืืฉ ืืืืง ืืฉืืื ืื ืืชืชื ืืฉืืืืช ืืื ืืื ืืืืขื.
|
non_infrastructure
|
ืืืืืฉ ืงืืืื ื ืืืืืฉ ืงืืืื ื ืืคื ืืืจืืฉืืช ืืืืคืืขืืช ืืืจืกื ืืคืืื ืืืืืืฉ ืืืฉืง ืืฉืชืืฉ ื ืื ืืืฉืง ืืฉืชืืฉ ืืืขืืื ืขืืืจ ืื ืกืื ืืฉืชืืฉ ืืขืื ืืฉ ืืืืง ืืฉืืื ืื ืืชืชื ืืฉืืืืช ืืื ืืื ืืืืขื
| 0
|
24,758
| 17,694,295,504
|
IssuesEvent
|
2021-08-24 13:45:32
|
gnosis/safe-ios
|
https://api.github.com/repos/gnosis/safe-ios
|
closed
|
Release 3.0.0
|
infrastructure
|
- [x] Create a release task in GitHub using the โNew Releaseโ template.
- [x] Create and push the release branch
```
git checkout main -b release/3.0.0
git push -u origin release/3.0.0
```
- [x] Marketing version is updated (3.0.0)
```
agvtool new-marketing-version 3.0.0
```
- [x] Notify QA
- [x] QA approved release candidate build
- [x] Product Owner approved submission
**AFTER PRODUCT OWNER APPROVAL**
- [x] Update screenshots in the App Store
- [x] Submit to the App Store Review with developer approval for distribution
- [x] Notify the team that release was submitted using the template below:
```
@here Hi everyone! We have submitted new iOS app v3.0.0 for review to the App Store.
```
- [x] Create a new release in GitHub with release notes. This will create a tag. The tag should be in a format v3.0.0
#### Download DSYMs manually
- [x] dSYMs are downloaded from AppStoreConnect and uploaded to Firebase Crashlytics.
```
# For the Multisig app (App Store version):
> ./bin/upload-symbols \
-gsp Multisig/Cross-layer/Analytics/Firebase/GoogleService-Info.Production.Mainnet.plist \
-p ios /path/to/dSYMs
```
#### Or download DSYMs with the script
- Install fastlane with `gem install fastlane --verbose`
- Set up the `fastlane` directory with configuraiton (ask team member to help). Do not commit the directory to the repository.
- Change the build version and build number in the `fastlane/upload_dsyms.sh` file
- Run the script `sh fastlane/upload_dsyms.sh`
#### Finally
- [x] Release the app when it is approved by the App Store Review team (do not release on Thu/Fri). Notify the team using the following template:
```
@here Hi everyone! We have released the iOS app v3.0.0 to the App Store and it will soon be available for download.
```
- [x] Merge the release branch to master branch via new pull-request
|
1.0
|
Release 3.0.0 - - [x] Create a release task in GitHub using the โNew Releaseโ template.
- [x] Create and push the release branch
```
git checkout main -b release/3.0.0
git push -u origin release/3.0.0
```
- [x] Marketing version is updated (3.0.0)
```
agvtool new-marketing-version 3.0.0
```
- [x] Notify QA
- [x] QA approved release candidate build
- [x] Product Owner approved submission
**AFTER PRODUCT OWNER APPROVAL**
- [x] Update screenshots in the App Store
- [x] Submit to the App Store Review with developer approval for distribution
- [x] Notify the team that release was submitted using the template below:
```
@here Hi everyone! We have submitted new iOS app v3.0.0 for review to the App Store.
```
- [x] Create a new release in GitHub with release notes. This will create a tag. The tag should be in a format v3.0.0
#### Download DSYMs manually
- [x] dSYMs are downloaded from AppStoreConnect and uploaded to Firebase Crashlytics.
```
# For the Multisig app (App Store version):
> ./bin/upload-symbols \
-gsp Multisig/Cross-layer/Analytics/Firebase/GoogleService-Info.Production.Mainnet.plist \
-p ios /path/to/dSYMs
```
#### Or download DSYMs with the script
- Install fastlane with `gem install fastlane --verbose`
- Set up the `fastlane` directory with configuraiton (ask team member to help). Do not commit the directory to the repository.
- Change the build version and build number in the `fastlane/upload_dsyms.sh` file
- Run the script `sh fastlane/upload_dsyms.sh`
#### Finally
- [x] Release the app when it is approved by the App Store Review team (do not release on Thu/Fri). Notify the team using the following template:
```
@here Hi everyone! We have released the iOS app v3.0.0 to the App Store and it will soon be available for download.
```
- [x] Merge the release branch to master branch via new pull-request
|
infrastructure
|
release create a release task in github using the โnew releaseโ template create and push the release branch git checkout main b release git push u origin release marketing version is updated agvtool new marketing version notify qa qa approved release candidate build product owner approved submission after product owner approval update screenshots in the app store submit to the app store review with developer approval for distribution notify the team that release was submitted using the template below here hi everyone we have submitted new ios app for review to the app store create a new release in github with release notes this will create a tag the tag should be in a format download dsyms manually dsyms are downloaded from appstoreconnect and uploaded to firebase crashlytics for the multisig app app store version bin upload symbols gsp multisig cross layer analytics firebase googleservice info production mainnet plist p ios path to dsyms or download dsyms with the script install fastlane with gem install fastlane verbose set up the fastlane directory with configuraiton ask team member to help do not commit the directory to the repository change the build version and build number in the fastlane upload dsyms sh file run the script sh fastlane upload dsyms sh finally release the app when it is approved by the app store review team do not release on thu fri notify the team using the following template here hi everyone we have released the ios app to the app store and it will soon be available for download merge the release branch to master branch via new pull request
| 1
|
94,817
| 16,021,737,389
|
IssuesEvent
|
2021-04-21 01:09:54
|
TIBCOSoftware/vscode-extension-tci
|
https://api.github.com/repos/TIBCOSoftware/vscode-extension-tci
|
opened
|
CVE-2021-23337 (High) detected in lodash-4.17.11.tgz
|
security vulnerability
|
## CVE-2021-23337 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.11.tgz</b></p></summary>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz</a></p>
<p>Path to dependency file: /vscode-extension-tci/package.json</p>
<p>Path to vulnerable library: vscode-extension-tci/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- archiver-2.1.1.tgz (Root Library)
- :x: **lodash-4.17.11.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function.
<p>Publish Date: 2021-02-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337>CVE-2021-23337</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c">https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c</a></p>
<p>Release Date: 2021-02-15</p>
<p>Fix Resolution: lodash - 4.17.21</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"4.17.11","packageFilePaths":["/vscode-extension-tci/package.json"],"isTransitiveDependency":true,"dependencyTree":"archiver:2.1.1;lodash:4.17.11","isMinimumFixVersionAvailable":true,"minimumFixVersion":"lodash - 4.17.21"}],"baseBranches":[],"vulnerabilityIdentifier":"CVE-2021-23337","vulnerabilityDetails":"Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337","cvss3Severity":"high","cvss3Score":"7.2","cvss3Metrics":{"A":"High","AC":"Low","PR":"High","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2021-23337 (High) detected in lodash-4.17.11.tgz - ## CVE-2021-23337 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.11.tgz</b></p></summary>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz</a></p>
<p>Path to dependency file: /vscode-extension-tci/package.json</p>
<p>Path to vulnerable library: vscode-extension-tci/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- archiver-2.1.1.tgz (Root Library)
- :x: **lodash-4.17.11.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function.
<p>Publish Date: 2021-02-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337>CVE-2021-23337</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: High
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c">https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c</a></p>
<p>Release Date: 2021-02-15</p>
<p>Fix Resolution: lodash - 4.17.21</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"4.17.11","packageFilePaths":["/vscode-extension-tci/package.json"],"isTransitiveDependency":true,"dependencyTree":"archiver:2.1.1;lodash:4.17.11","isMinimumFixVersionAvailable":true,"minimumFixVersion":"lodash - 4.17.21"}],"baseBranches":[],"vulnerabilityIdentifier":"CVE-2021-23337","vulnerabilityDetails":"Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337","cvss3Severity":"high","cvss3Score":"7.2","cvss3Metrics":{"A":"High","AC":"Low","PR":"High","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_infrastructure
|
cve high detected in lodash tgz cve high severity vulnerability vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file vscode extension tci package json path to vulnerable library vscode extension tci node modules lodash package json dependency hierarchy archiver tgz root library x lodash tgz vulnerable library vulnerability details lodash versions prior to are vulnerable to command injection via the template function publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree archiver lodash isminimumfixversionavailable true minimumfixversion lodash basebranches vulnerabilityidentifier cve vulnerabilitydetails lodash versions prior to are vulnerable to command injection via the template function vulnerabilityurl
| 0
|
17,537
| 12,419,977,010
|
IssuesEvent
|
2020-05-23 09:27:52
|
libero/reviewer
|
https://api.github.com/repos/libero/reviewer
|
closed
|
Journal Login uses a static submit url, ignores request
|
Infrastructure
|
Whenever I use it to login on the new[ reviewer-staging](https://continuumtest--cdn-journal.elifesciences.org/) I end up being redirected to the soon to be replaced [staging](https://libero-reviewer--staging.elifesciences.org/login).
Looking at the network traffic the request sent when I click _Login_ has `Referer: https://reviewer--staging--test-cluster.elifesciences.org/login` set, so I'd expect `Location` to be set accordingly in the response, but it just points at the old staging.
Turn out the url is set statically when a journal instance is created:
https://github.com/elifesciences/builder-configuration/blob/3ab43524d163b1d69e09e5a2a8e194ec6afd0cc4/pillar/environment-continuumtest-public.sls
https://github.com/elifesciences/journal/blob/fd17edfb7c47f54b9c4d8cfb7d4f81ce636091b5/src/Controller/SubmitController.php
|
1.0
|
Journal Login uses a static submit url, ignores request - Whenever I use it to login on the new[ reviewer-staging](https://continuumtest--cdn-journal.elifesciences.org/) I end up being redirected to the soon to be replaced [staging](https://libero-reviewer--staging.elifesciences.org/login).
Looking at the network traffic the request sent when I click _Login_ has `Referer: https://reviewer--staging--test-cluster.elifesciences.org/login` set, so I'd expect `Location` to be set accordingly in the response, but it just points at the old staging.
Turn out the url is set statically when a journal instance is created:
https://github.com/elifesciences/builder-configuration/blob/3ab43524d163b1d69e09e5a2a8e194ec6afd0cc4/pillar/environment-continuumtest-public.sls
https://github.com/elifesciences/journal/blob/fd17edfb7c47f54b9c4d8cfb7d4f81ce636091b5/src/Controller/SubmitController.php
|
infrastructure
|
journal login uses a static submit url ignores request whenever i use it to login on the new i end up being redirected to the soon to be replaced looking at the network traffic the request sent when i click login has referer set so i d expect location to be set accordingly in the response but it just points at the old staging turn out the url is set statically when a journal instance is created
| 1
|
31,198
| 25,408,425,391
|
IssuesEvent
|
2022-11-22 16:55:35
|
SonarSource/sonarlint-visualstudio
|
https://api.github.com/repos/SonarSource/sonarlint-visualstudio
|
opened
|
[Infra] Auto-regenerate the list of Sonar Roslyn rules that can be suppressed on build
|
Infrastructure
|
### Description
The list of issues C#/VB rules that can be suppress in Connected Mode is generated using a [T4 template](https://github.com/SonarSource/sonarlint-visualstudio/blob/69de0102f4155f7f45aafebbc41b0b79f2958af2/src/Roslyn.Suppressions/Roslyn.Suppressions/SupportedSuppressionBuilder.tt#L1).
At present the code is only regenerated manually i.e. we have to remember to do it each time we update the Roslyn analyzers.
We should use the T4 build integration to cause the file to be updated automatically on build.
|
1.0
|
[Infra] Auto-regenerate the list of Sonar Roslyn rules that can be suppressed on build - ### Description
The list of issues C#/VB rules that can be suppress in Connected Mode is generated using a [T4 template](https://github.com/SonarSource/sonarlint-visualstudio/blob/69de0102f4155f7f45aafebbc41b0b79f2958af2/src/Roslyn.Suppressions/Roslyn.Suppressions/SupportedSuppressionBuilder.tt#L1).
At present the code is only regenerated manually i.e. we have to remember to do it each time we update the Roslyn analyzers.
We should use the T4 build integration to cause the file to be updated automatically on build.
|
infrastructure
|
auto regenerate the list of sonar roslyn rules that can be suppressed on build description the list of issues c vb rules that can be suppress in connected mode is generated using a at present the code is only regenerated manually i e we have to remember to do it each time we update the roslyn analyzers we should use the build integration to cause the file to be updated automatically on build
| 1
|
12,596
| 9,875,146,040
|
IssuesEvent
|
2019-06-23 09:04:50
|
OpenCHS/openchs-product
|
https://api.github.com/repos/OpenCHS/openchs-product
|
closed
|
Spike for metabase
|
0.5 Complete Infrastructure/other Must Story
|
Spike ability to support
Multitenancy
Reporting from json structure
ย
Results
Multitenancy can be achieved by using row level security.
Create multiple connections, each with a different user. Give database privileges per tenant.ย
ย
A few reports from json structure attached with this story.
|
1.0
|
Spike for metabase - Spike ability to support
Multitenancy
Reporting from json structure
ย
Results
Multitenancy can be achieved by using row level security.
Create multiple connections, each with a different user. Give database privileges per tenant.ย
ย
A few reports from json structure attached with this story.
|
infrastructure
|
spike for metabase spike ability to support multitenancy reporting from json structure ย results multitenancy can be achieved by using row level security create multiple connections each with a different user give database privileges per tenant ย ย a few reports from json structure attached with this story
| 1
|
123,558
| 12,201,696,756
|
IssuesEvent
|
2020-04-30 07:42:31
|
react-hook-form/react-hook-form
|
https://api.github.com/repos/react-hook-form/react-hook-form
|
closed
|
defaultValues for radio inputs does not set formState.isValid
|
improve documentation
|
**Describe the bug**
When using the `defaultValues` param of the `useForm` function, one expects all fields passed a value to be filled in and for the form to be valid if all required fields are thus filled.
When using defaultValues with text inputs, this is the case: `formState.isValid` is true. However when using the defaultValues of radio inputs: formState.isValid returns false, even though a radio option is showing as selected.
**To Reproduce**
Steps to reproduce the behavior:
1. Visit [this sandbox](https://codesandbox.io/s/react-hook-radio-default-vales-vkqet?file=/src/App.js)
2. Reload the page
3. The `formState` is not valid, but defaultValues has selected a radio option in the only required field
**Codesandbox link (Required)**
https://codesandbox.io/s/react-hook-radio-default-vales-vkqet?file=/src/App.js
**Expected behavior**
`defaultValues` causes `formState.isValid` to be true if all required fields have values.
**Desktop (please complete the following information):**
- OS: Mac OSX
- Browser: Chrome
|
1.0
|
defaultValues for radio inputs does not set formState.isValid - **Describe the bug**
When using the `defaultValues` param of the `useForm` function, one expects all fields passed a value to be filled in and for the form to be valid if all required fields are thus filled.
When using defaultValues with text inputs, this is the case: `formState.isValid` is true. However when using the defaultValues of radio inputs: formState.isValid returns false, even though a radio option is showing as selected.
**To Reproduce**
Steps to reproduce the behavior:
1. Visit [this sandbox](https://codesandbox.io/s/react-hook-radio-default-vales-vkqet?file=/src/App.js)
2. Reload the page
3. The `formState` is not valid, but defaultValues has selected a radio option in the only required field
**Codesandbox link (Required)**
https://codesandbox.io/s/react-hook-radio-default-vales-vkqet?file=/src/App.js
**Expected behavior**
`defaultValues` causes `formState.isValid` to be true if all required fields have values.
**Desktop (please complete the following information):**
- OS: Mac OSX
- Browser: Chrome
|
non_infrastructure
|
defaultvalues for radio inputs does not set formstate isvalid describe the bug when using the defaultvalues param of the useform function one expects all fields passed a value to be filled in and for the form to be valid if all required fields are thus filled when using defaultvalues with text inputs this is the case formstate isvalid is true however when using the defaultvalues of radio inputs formstate isvalid returns false even though a radio option is showing as selected to reproduce steps to reproduce the behavior visit reload the page the formstate is not valid but defaultvalues has selected a radio option in the only required field codesandbox link required expected behavior defaultvalues causes formstate isvalid to be true if all required fields have values desktop please complete the following information os mac osx browser chrome
| 0
|
14,284
| 10,736,684,907
|
IssuesEvent
|
2019-10-29 11:26:49
|
MeteoSwiss-APN/dawn
|
https://api.github.com/repos/MeteoSwiss-APN/dawn
|
closed
|
Add data to IIR AST: 1. Accesses pair
|
infrastructure
|
DoD: replace the accesses pair to include the new AST node with data
|
1.0
|
Add data to IIR AST: 1. Accesses pair - DoD: replace the accesses pair to include the new AST node with data
|
infrastructure
|
add data to iir ast accesses pair dod replace the accesses pair to include the new ast node with data
| 1
|
84,544
| 7,925,253,760
|
IssuesEvent
|
2018-07-05 19:56:36
|
ONRR/doi-extractives-data
|
https://api.github.com/repos/ONRR/doi-extractives-data
|
closed
|
Updated CY Federal Production
|
workflow:testing
|
- [x] Check file format and known data issues
- [x] Update visualizations
|
1.0
|
Updated CY Federal Production - - [x] Check file format and known data issues
- [x] Update visualizations
|
non_infrastructure
|
updated cy federal production check file format and known data issues update visualizations
| 0
|
4,437
| 5,075,476,634
|
IssuesEvent
|
2016-12-27 19:52:49
|
dotnet/corefx
|
https://api.github.com/repos/dotnet/corefx
|
closed
|
Remove SNI workaround when we have a new TFS build
|
area-Infrastructure
|
https://github.com/dotnet/corefx/pull/12376/commits/c7ba063079560f0a6c1057a62da51a4ef53de629
We should remove this once we get a new TFS build. /cc @stephentoub
|
1.0
|
Remove SNI workaround when we have a new TFS build - https://github.com/dotnet/corefx/pull/12376/commits/c7ba063079560f0a6c1057a62da51a4ef53de629
We should remove this once we get a new TFS build. /cc @stephentoub
|
infrastructure
|
remove sni workaround when we have a new tfs build we should remove this once we get a new tfs build cc stephentoub
| 1
|
10,802
| 8,728,752,085
|
IssuesEvent
|
2018-12-10 18:16:10
|
dotnet/corefx
|
https://api.github.com/repos/dotnet/corefx
|
opened
|
Update code coverage template
|
area-Infrastructure
|
The current template does not contain all features that come with ReportGenerator. Also it contains a zip link at the top right which indicates that it is custom tailored.
@mmitche @MattGal is the htmlPublisher using a static template? how can we update it?
cc @danmosemsft
|
1.0
|
Update code coverage template - The current template does not contain all features that come with ReportGenerator. Also it contains a zip link at the top right which indicates that it is custom tailored.
@mmitche @MattGal is the htmlPublisher using a static template? how can we update it?
cc @danmosemsft
|
infrastructure
|
update code coverage template the current template does not contain all features that come with reportgenerator also it contains a zip link at the top right which indicates that it is custom tailored mmitche mattgal is the htmlpublisher using a static template how can we update it cc danmosemsft
| 1
|
64,839
| 14,686,226,361
|
IssuesEvent
|
2021-01-01 13:53:17
|
LalithK90/Welfare
|
https://api.github.com/repos/LalithK90/Welfare
|
opened
|
CVE-2020-25638 (High) detected in hibernate-core-5.4.10.Final.jar
|
security vulnerability
|
## CVE-2020-25638 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hibernate-core-5.4.10.Final.jar</b></p></summary>
<p>Hibernate's core ORM functionality</p>
<p>Library home page: <a href="http://hibernate.org/orm">http://hibernate.org/orm</a></p>
<p>Path to dependency file: Welfare/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.hibernate/hibernate-core/5.4.10.Final/365690f87b040dc5d22cc24a4daff76d1cffef23/hibernate-core-5.4.10.Final.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-data-jpa-2.2.4.RELEASE.jar (Root Library)
- :x: **hibernate-core-5.4.10.Final.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in hibernate-core in versions prior to and including 5.4.23.Final. A SQL injection in the implementation of the JPA Criteria API can permit unsanitized literals when a literal is used in the SQL comments of the query. This flaw could allow an attacker to access unauthorized information or possibly conduct further attacks. The highest threat from this vulnerability is to data confidentiality and integrity.
<p>Publish Date: 2020-12-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-25638>CVE-2020-25638</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://in.relation.to/2020/11/19/hibernate-orm-5424-final-release/">https://in.relation.to/2020/11/19/hibernate-orm-5424-final-release/</a></p>
<p>Release Date: 2020-12-02</p>
<p>Fix Resolution: org.hibernate:hibernate-core:5.3.20.Final,5.4.24.Final</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2020-25638 (High) detected in hibernate-core-5.4.10.Final.jar - ## CVE-2020-25638 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hibernate-core-5.4.10.Final.jar</b></p></summary>
<p>Hibernate's core ORM functionality</p>
<p>Library home page: <a href="http://hibernate.org/orm">http://hibernate.org/orm</a></p>
<p>Path to dependency file: Welfare/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.hibernate/hibernate-core/5.4.10.Final/365690f87b040dc5d22cc24a4daff76d1cffef23/hibernate-core-5.4.10.Final.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-data-jpa-2.2.4.RELEASE.jar (Root Library)
- :x: **hibernate-core-5.4.10.Final.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in hibernate-core in versions prior to and including 5.4.23.Final. A SQL injection in the implementation of the JPA Criteria API can permit unsanitized literals when a literal is used in the SQL comments of the query. This flaw could allow an attacker to access unauthorized information or possibly conduct further attacks. The highest threat from this vulnerability is to data confidentiality and integrity.
<p>Publish Date: 2020-12-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-25638>CVE-2020-25638</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://in.relation.to/2020/11/19/hibernate-orm-5424-final-release/">https://in.relation.to/2020/11/19/hibernate-orm-5424-final-release/</a></p>
<p>Release Date: 2020-12-02</p>
<p>Fix Resolution: org.hibernate:hibernate-core:5.3.20.Final,5.4.24.Final</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_infrastructure
|
cve high detected in hibernate core final jar cve high severity vulnerability vulnerable library hibernate core final jar hibernate s core orm functionality library home page a href path to dependency file welfare build gradle path to vulnerable library home wss scanner gradle caches modules files org hibernate hibernate core final hibernate core final jar dependency hierarchy spring boot starter data jpa release jar root library x hibernate core final jar vulnerable library vulnerability details a flaw was found in hibernate core in versions prior to and including final a sql injection in the implementation of the jpa criteria api can permit unsanitized literals when a literal is used in the sql comments of the query this flaw could allow an attacker to access unauthorized information or possibly conduct further attacks the highest threat from this vulnerability is to data confidentiality and integrity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org hibernate hibernate core final final step up your open source security game with whitesource
| 0
|
529,860
| 15,396,703,559
|
IssuesEvent
|
2021-03-03 21:01:43
|
dell/karavi-observability
|
https://api.github.com/repos/dell/karavi-observability
|
closed
|
[BUG]: Deprecation warning messages for CSINode and StorageClass in karavi-metrics-powerflex
|
priority/high release-found/0.2.0 source/karavi-metrics-powerflex type/bug
|
**Describe the bug**
Using the latest from the main branch deployed in k8s v1.20.0, the logs for karavi-metrics-powerflex are showing the following deprecation warning messages:
```
warnings.go:70] storage.k8s.io/v1beta1 CSINode is deprecated in v1.17+, unavailable in v1.22+; use storage.k8s.io/v1 CSINode
warnings.go:70] storage.k8s.io/v1beta1 StorageClass is deprecated in v1.19+, unavailable in v1.22+; use storage.k8s.io/v1 StorageClass
```
**To Reproduce**
Steps to reproduce the behavior:
1. Build and deploy an image from the main branch for karavi-metrics-powerflex
2. View the logs for the deployment/pod
**Expected behavior**
These deprecation warnings should not be showing up in the logs.
**Screenshots**
N/A
**Logs**
N/A
**System Information (please complete the following information):**
- OS/Version: [e.g. RHEL 7.6]
- Kubernetes Version [e.g. 1.18]: v1.20.0
- karavi-metrics-powerflex Version: [e.g. 0.1.0-pre-release]
- karavi-topology Version: [e.g. 0.1.0-pre-release]
- Prometheus Version: [e.g. 2.22.0]
- Grafana Version: [e.g. 7.3.2]
- Grafana Dashboard Title(s): [e.g. Storage Pool and Storage Class Capacity]
- Grafana Datatable Panel plugin configuration (can be provided as a screenshot)
- Grafana Prometheus Datasource configuration (can be provided as a screenshot)
- Grafana Pie Chart Panel plugin configuration (can be provided as a screenshot)
- Grafana SimpleJson Datasource configuration (can be provided as a screenshot)
- Additional Information...
**Additional context**
N/A
|
1.0
|
[BUG]: Deprecation warning messages for CSINode and StorageClass in karavi-metrics-powerflex - **Describe the bug**
Using the latest from the main branch deployed in k8s v1.20.0, the logs for karavi-metrics-powerflex are showing the following deprecation warning messages:
```
warnings.go:70] storage.k8s.io/v1beta1 CSINode is deprecated in v1.17+, unavailable in v1.22+; use storage.k8s.io/v1 CSINode
warnings.go:70] storage.k8s.io/v1beta1 StorageClass is deprecated in v1.19+, unavailable in v1.22+; use storage.k8s.io/v1 StorageClass
```
**To Reproduce**
Steps to reproduce the behavior:
1. Build and deploy an image from the main branch for karavi-metrics-powerflex
2. View the logs for the deployment/pod
**Expected behavior**
These deprecation warnings should not be showing up in the logs.
**Screenshots**
N/A
**Logs**
N/A
**System Information (please complete the following information):**
- OS/Version: [e.g. RHEL 7.6]
- Kubernetes Version [e.g. 1.18]: v1.20.0
- karavi-metrics-powerflex Version: [e.g. 0.1.0-pre-release]
- karavi-topology Version: [e.g. 0.1.0-pre-release]
- Prometheus Version: [e.g. 2.22.0]
- Grafana Version: [e.g. 7.3.2]
- Grafana Dashboard Title(s): [e.g. Storage Pool and Storage Class Capacity]
- Grafana Datatable Panel plugin configuration (can be provided as a screenshot)
- Grafana Prometheus Datasource configuration (can be provided as a screenshot)
- Grafana Pie Chart Panel plugin configuration (can be provided as a screenshot)
- Grafana SimpleJson Datasource configuration (can be provided as a screenshot)
- Additional Information...
**Additional context**
N/A
|
non_infrastructure
|
deprecation warning messages for csinode and storageclass in karavi metrics powerflex describe the bug using the latest from the main branch deployed in the logs for karavi metrics powerflex are showing the following deprecation warning messages warnings go storage io csinode is deprecated in unavailable in use storage io csinode warnings go storage io storageclass is deprecated in unavailable in use storage io storageclass to reproduce steps to reproduce the behavior build and deploy an image from the main branch for karavi metrics powerflex view the logs for the deployment pod expected behavior these deprecation warnings should not be showing up in the logs screenshots n a logs n a system information please complete the following information os version kubernetes version karavi metrics powerflex version karavi topology version prometheus version grafana version grafana dashboard title s grafana datatable panel plugin configuration can be provided as a screenshot grafana prometheus datasource configuration can be provided as a screenshot grafana pie chart panel plugin configuration can be provided as a screenshot grafana simplejson datasource configuration can be provided as a screenshot additional information additional context n a
| 0
|
20,139
| 13,690,257,362
|
IssuesEvent
|
2020-09-30 14:10:20
|
gather-community/gather
|
https://api.github.com/repos/gather-community/gather
|
closed
|
Setting up Discourse instance
|
area:discourse type:infrastructure
|
_Originally created by **Tom Smyth** at **2018-03-30 15:42**, migrated from [redmine-#8226](https://redmine.sassafras.coop/issues/8226)_
|
1.0
|
Setting up Discourse instance - _Originally created by **Tom Smyth** at **2018-03-30 15:42**, migrated from [redmine-#8226](https://redmine.sassafras.coop/issues/8226)_
|
infrastructure
|
setting up discourse instance originally created by tom smyth at migrated from
| 1
|
279,967
| 30,787,335,408
|
IssuesEvent
|
2023-07-31 14:02:15
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
opened
|
[API Proposal]: Provide an implementation AesGcm in Microsoft.Bcl.Cryptography
|
api-suggestion area-System.Security
|
### Background and motivation
The `AesGcm` class was introduced in to .NET in .NET Core 3.1. AES-GCM is an AEAD that is used in many protocols and specifications. Its broad use in in those specs makes it desirable or mandatory to support them.
For .NET Framework, there is no AES-GCM implementation that is readily available. Developers that want AES-GCM either have to p/invoke to Windows or find some other alternative implementation.
This is a proposal to add `AesGcm` to [Microsoft.Bcl.Cryptography][mbc].
1. When the package is build for .NET (Core), it will type-forward in to the shared runtime implementation.
2. When the package is build for .NET Framework, it will carry the required API surface and p/invokes similar to `SP800108HmacCounterKdf`
3. When the package is build for .NET Standard, it will behave similar to .NET Framework, however will perform a runtime check that the platform is Windows.
This means that the package will only provide an implementation on Windows. This is largely okay - developers on non-Windows platforms already have this in .NET itself and will benefit from the type forwards.
[mbc]: https://www.nuget.org/packages/Microsoft.Bcl.Cryptography/
### API Proposal
To provide the best experience, I think the `ref` needs to be split between .NET, .NET Framework, and .NET Standard.
This is because of the constructors that were [replaced in .NET 8](https://github.com/dotnet/runtime/issues/71366). For .NET Framework and .NET Standard, it would be ideal that we provide them with the same improved constructors since the new constructors are meant to mitigate misuse.
For .NET, we need the API shape to make between the respective versions in order for the type forward to work, otherwise we'll get MissingMemberExceptions when the type forwards the new constructors.
As those versions of .NET age out of support, the ref-split can be simplified to just .NET Framework / Standard and .NET.
### .NET Framework and .NET Standard
```csharp
public sealed partial class AesGcm : System.IDisposable
{
// NOT included because they are obsolete
// public AesGcm(byte[] key) { }
// public AesGcm(System.ReadOnlySpan<byte> key) { }
public AesGcm(byte[] key, int tagSizeInBytes) { }
public AesGcm(System.ReadOnlySpan<byte> key, int tagSizeInBytes) { }
public static bool IsSupported { get { throw null; } }
public static System.Security.Cryptography.KeySizes NonceByteSizes { get { throw null; } }
public static System.Security.Cryptography.KeySizes TagByteSizes { get { throw null; } }
public int? TagSizeInBytes { get { throw null; } }
public void Decrypt(byte[] nonce, byte[] ciphertext, byte[] tag, byte[] plaintext, byte[]? associatedData = null) { }
public void Decrypt(System.ReadOnlySpan<byte> nonce, System.ReadOnlySpan<byte> ciphertext, System.ReadOnlySpan<byte> tag, System.Span<byte> plaintext, System.ReadOnlySpan<byte> associatedData = default(System.ReadOnlySpan<byte>)) { }
public void Dispose() { }
public void Encrypt(byte[] nonce, byte[] plaintext, byte[] ciphertext, byte[] tag, byte[]? associatedData = null) { }
public void Encrypt(System.ReadOnlySpan<byte> nonce, System.ReadOnlySpan<byte> plaintext, System.Span<byte> ciphertext, System.Span<byte> tag, System.ReadOnlySpan<byte> associatedData = default(System.ReadOnlySpan<byte>)) { }
}
```
### .NET 6, 7 type forward
```csharp
// Matches ref from .NET 6 and 7.
public sealed partial class AesGcm : System.IDisposable
{
public AesGcm(byte[] key) { }
public AesGcm(System.ReadOnlySpan<byte> key) { }
public static bool IsSupported { get { throw null; } }
public static System.Security.Cryptography.KeySizes NonceByteSizes { get { throw null; } }
public static System.Security.Cryptography.KeySizes TagByteSizes { get { throw null; } }
public void Decrypt(byte[] nonce, byte[] ciphertext, byte[] tag, byte[] plaintext, byte[]? associatedData = null) { }
public void Decrypt(System.ReadOnlySpan<byte> nonce, System.ReadOnlySpan<byte> ciphertext, System.ReadOnlySpan<byte> tag, System.Span<byte> plaintext, System.ReadOnlySpan<byte> associatedData = default(System.ReadOnlySpan<byte>)) { }
public void Dispose() { }
public void Encrypt(byte[] nonce, byte[] plaintext, byte[] ciphertext, byte[] tag, byte[]? associatedData = null) { }
public void Encrypt(System.ReadOnlySpan<byte> nonce, System.ReadOnlySpan<byte> plaintext, System.Span<byte> ciphertext, System.Span<byte> tag, System.ReadOnlySpan<byte> associatedData = default(System.ReadOnlySpan<byte>)) { }
}
```
### .NET 8 type forward
```csharp
// Matches ref for .NET 8.
public sealed partial class AesGcm : System.IDisposable
{
[System.ObsoleteAttribute("AesGcm should indicate the required tag size for encryption and decryption. Use a constructor that accepts the tag size.", DiagnosticId="SYSLIB0053", UrlFormat="https://aka.ms/dotnet-warnings/{0}")]
public AesGcm(byte[] key) { }
public AesGcm(byte[] key, int tagSizeInBytes) { }
[System.ObsoleteAttribute("AesGcm should indicate the required tag size for encryption and decryption. Use a constructor that accepts the tag size.", DiagnosticId="SYSLIB0053", UrlFormat="https://aka.ms/dotnet-warnings/{0}")]
public AesGcm(System.ReadOnlySpan<byte> key) { }
public AesGcm(System.ReadOnlySpan<byte> key, int tagSizeInBytes) { }
public static bool IsSupported { get { throw null; } }
public static System.Security.Cryptography.KeySizes NonceByteSizes { get { throw null; } }
public static System.Security.Cryptography.KeySizes TagByteSizes { get { throw null; } }
public int? TagSizeInBytes { get { throw null; } }
public void Decrypt(byte[] nonce, byte[] ciphertext, byte[] tag, byte[] plaintext, byte[]? associatedData = null) { }
public void Decrypt(System.ReadOnlySpan<byte> nonce, System.ReadOnlySpan<byte> ciphertext, System.ReadOnlySpan<byte> tag, System.Span<byte> plaintext, System.ReadOnlySpan<byte> associatedData = default(System.ReadOnlySpan<byte>)) { }
public void Dispose() { }
public void Encrypt(byte[] nonce, byte[] plaintext, byte[] ciphertext, byte[] tag, byte[]? associatedData = null) { }
public void Encrypt(System.ReadOnlySpan<byte> nonce, System.ReadOnlySpan<byte> plaintext, System.Span<byte> ciphertext, System.Span<byte> tag, System.ReadOnlySpan<byte> associatedData = default(System.ReadOnlySpan<byte>)) { }
}
```
### API Usage
Same as .NET.
### Alternative Designs
1. This proposal leaves out `AesCcm`. This omission is purely due to no known demand for AES-CCM in downlevel platforms, whereas AES-GCM has demonstrated considerable demand. We can include AesCcm for completeness if desired.
2. The ref-split is slightly complicated, but will improve as .NET 6 and 7 age out. The "simplest" thing to do is to outbox only the ref from .NET 6. However I think this is incorrect because it would mean outboxing constructors that are obsolete. There is no reason that .NET Framework needs the obsolete constructors. We could include both the new and obsolete constructors for .NET Framework if that makes multi-targeting easier.
### Risks
This carries the similar risks of providing any API as a package. It carries maintenance burden.
|
True
|
[API Proposal]: Provide an implementation AesGcm in Microsoft.Bcl.Cryptography - ### Background and motivation
The `AesGcm` class was introduced in to .NET in .NET Core 3.1. AES-GCM is an AEAD that is used in many protocols and specifications. Its broad use in in those specs makes it desirable or mandatory to support them.
For .NET Framework, there is no AES-GCM implementation that is readily available. Developers that want AES-GCM either have to p/invoke to Windows or find some other alternative implementation.
This is a proposal to add `AesGcm` to [Microsoft.Bcl.Cryptography][mbc].
1. When the package is build for .NET (Core), it will type-forward in to the shared runtime implementation.
2. When the package is build for .NET Framework, it will carry the required API surface and p/invokes similar to `SP800108HmacCounterKdf`
3. When the package is build for .NET Standard, it will behave similar to .NET Framework, however will perform a runtime check that the platform is Windows.
This means that the package will only provide an implementation on Windows. This is largely okay - developers on non-Windows platforms already have this in .NET itself and will benefit from the type forwards.
[mbc]: https://www.nuget.org/packages/Microsoft.Bcl.Cryptography/
### API Proposal
To provide the best experience, I think the `ref` needs to be split between .NET, .NET Framework, and .NET Standard.
This is because of the constructors that were [replaced in .NET 8](https://github.com/dotnet/runtime/issues/71366). For .NET Framework and .NET Standard, it would be ideal that we provide them with the same improved constructors since the new constructors are meant to mitigate misuse.
For .NET, we need the API shape to make between the respective versions in order for the type forward to work, otherwise we'll get MissingMemberExceptions when the type forwards the new constructors.
As those versions of .NET age out of support, the ref-split can be simplified to just .NET Framework / Standard and .NET.
### .NET Framework and .NET Standard
```csharp
public sealed partial class AesGcm : System.IDisposable
{
// NOT included because they are obsolete
// public AesGcm(byte[] key) { }
// public AesGcm(System.ReadOnlySpan<byte> key) { }
public AesGcm(byte[] key, int tagSizeInBytes) { }
public AesGcm(System.ReadOnlySpan<byte> key, int tagSizeInBytes) { }
public static bool IsSupported { get { throw null; } }
public static System.Security.Cryptography.KeySizes NonceByteSizes { get { throw null; } }
public static System.Security.Cryptography.KeySizes TagByteSizes { get { throw null; } }
public int? TagSizeInBytes { get { throw null; } }
public void Decrypt(byte[] nonce, byte[] ciphertext, byte[] tag, byte[] plaintext, byte[]? associatedData = null) { }
public void Decrypt(System.ReadOnlySpan<byte> nonce, System.ReadOnlySpan<byte> ciphertext, System.ReadOnlySpan<byte> tag, System.Span<byte> plaintext, System.ReadOnlySpan<byte> associatedData = default(System.ReadOnlySpan<byte>)) { }
public void Dispose() { }
public void Encrypt(byte[] nonce, byte[] plaintext, byte[] ciphertext, byte[] tag, byte[]? associatedData = null) { }
public void Encrypt(System.ReadOnlySpan<byte> nonce, System.ReadOnlySpan<byte> plaintext, System.Span<byte> ciphertext, System.Span<byte> tag, System.ReadOnlySpan<byte> associatedData = default(System.ReadOnlySpan<byte>)) { }
}
```
### .NET 6, 7 type forward
```csharp
// Matches ref from .NET 6 and 7.
public sealed partial class AesGcm : System.IDisposable
{
public AesGcm(byte[] key) { }
public AesGcm(System.ReadOnlySpan<byte> key) { }
public static bool IsSupported { get { throw null; } }
public static System.Security.Cryptography.KeySizes NonceByteSizes { get { throw null; } }
public static System.Security.Cryptography.KeySizes TagByteSizes { get { throw null; } }
public void Decrypt(byte[] nonce, byte[] ciphertext, byte[] tag, byte[] plaintext, byte[]? associatedData = null) { }
public void Decrypt(System.ReadOnlySpan<byte> nonce, System.ReadOnlySpan<byte> ciphertext, System.ReadOnlySpan<byte> tag, System.Span<byte> plaintext, System.ReadOnlySpan<byte> associatedData = default(System.ReadOnlySpan<byte>)) { }
public void Dispose() { }
public void Encrypt(byte[] nonce, byte[] plaintext, byte[] ciphertext, byte[] tag, byte[]? associatedData = null) { }
public void Encrypt(System.ReadOnlySpan<byte> nonce, System.ReadOnlySpan<byte> plaintext, System.Span<byte> ciphertext, System.Span<byte> tag, System.ReadOnlySpan<byte> associatedData = default(System.ReadOnlySpan<byte>)) { }
}
```
### .NET 8 type forward
```csharp
// Matches ref for .NET 8.
public sealed partial class AesGcm : System.IDisposable
{
[System.ObsoleteAttribute("AesGcm should indicate the required tag size for encryption and decryption. Use a constructor that accepts the tag size.", DiagnosticId="SYSLIB0053", UrlFormat="https://aka.ms/dotnet-warnings/{0}")]
public AesGcm(byte[] key) { }
public AesGcm(byte[] key, int tagSizeInBytes) { }
[System.ObsoleteAttribute("AesGcm should indicate the required tag size for encryption and decryption. Use a constructor that accepts the tag size.", DiagnosticId="SYSLIB0053", UrlFormat="https://aka.ms/dotnet-warnings/{0}")]
public AesGcm(System.ReadOnlySpan<byte> key) { }
public AesGcm(System.ReadOnlySpan<byte> key, int tagSizeInBytes) { }
public static bool IsSupported { get { throw null; } }
public static System.Security.Cryptography.KeySizes NonceByteSizes { get { throw null; } }
public static System.Security.Cryptography.KeySizes TagByteSizes { get { throw null; } }
public int? TagSizeInBytes { get { throw null; } }
public void Decrypt(byte[] nonce, byte[] ciphertext, byte[] tag, byte[] plaintext, byte[]? associatedData = null) { }
public void Decrypt(System.ReadOnlySpan<byte> nonce, System.ReadOnlySpan<byte> ciphertext, System.ReadOnlySpan<byte> tag, System.Span<byte> plaintext, System.ReadOnlySpan<byte> associatedData = default(System.ReadOnlySpan<byte>)) { }
public void Dispose() { }
public void Encrypt(byte[] nonce, byte[] plaintext, byte[] ciphertext, byte[] tag, byte[]? associatedData = null) { }
public void Encrypt(System.ReadOnlySpan<byte> nonce, System.ReadOnlySpan<byte> plaintext, System.Span<byte> ciphertext, System.Span<byte> tag, System.ReadOnlySpan<byte> associatedData = default(System.ReadOnlySpan<byte>)) { }
}
```
### API Usage
Same as .NET.
### Alternative Designs
1. This proposal leaves out `AesCcm`. This omission is purely due to no known demand for AES-CCM in downlevel platforms, whereas AES-GCM has demonstrated considerable demand. We can include AesCcm for completeness if desired.
2. The ref-split is slightly complicated, but will improve as .NET 6 and 7 age out. The "simplest" thing to do is to outbox only the ref from .NET 6. However I think this is incorrect because it would mean outboxing constructors that are obsolete. There is no reason that .NET Framework needs the obsolete constructors. We could include both the new and obsolete constructors for .NET Framework if that makes multi-targeting easier.
### Risks
This carries the similar risks of providing any API as a package. It carries maintenance burden.
|
non_infrastructure
|
provide an implementation aesgcm in microsoft bcl cryptography background and motivation the aesgcm class was introduced in to net in net core aes gcm is an aead that is used in many protocols and specifications its broad use in in those specs makes it desirable or mandatory to support them for net framework there is no aes gcm implementation that is readily available developers that want aes gcm either have to p invoke to windows or find some other alternative implementation this is a proposal to add aesgcm to when the package is build for net core it will type forward in to the shared runtime implementation when the package is build for net framework it will carry the required api surface and p invokes similar to when the package is build for net standard it will behave similar to net framework however will perform a runtime check that the platform is windows this means that the package will only provide an implementation on windows this is largely okay developers on non windows platforms already have this in net itself and will benefit from the type forwards api proposal to provide the best experience i think the ref needs to be split between net net framework and net standard this is because of the constructors that were for net framework and net standard it would be ideal that we provide them with the same improved constructors since the new constructors are meant to mitigate misuse for net we need the api shape to make between the respective versions in order for the type forward to work otherwise we ll get missingmemberexceptions when the type forwards the new constructors as those versions of net age out of support the ref split can be simplified to just net framework standard and net net framework and net standard csharp public sealed partial class aesgcm system idisposable not included because they are obsolete public aesgcm byte key public aesgcm system readonlyspan key public aesgcm byte key int tagsizeinbytes public aesgcm system readonlyspan key int tagsizeinbytes public static bool issupported get throw null public static system security cryptography keysizes noncebytesizes get throw null public static system security cryptography keysizes tagbytesizes get throw null public int tagsizeinbytes get throw null public void decrypt byte nonce byte ciphertext byte tag byte plaintext byte associateddata null public void decrypt system readonlyspan nonce system readonlyspan ciphertext system readonlyspan tag system span plaintext system readonlyspan associateddata default system readonlyspan public void dispose public void encrypt byte nonce byte plaintext byte ciphertext byte tag byte associateddata null public void encrypt system readonlyspan nonce system readonlyspan plaintext system span ciphertext system span tag system readonlyspan associateddata default system readonlyspan net type forward csharp matches ref from net and public sealed partial class aesgcm system idisposable public aesgcm byte key public aesgcm system readonlyspan key public static bool issupported get throw null public static system security cryptography keysizes noncebytesizes get throw null public static system security cryptography keysizes tagbytesizes get throw null public void decrypt byte nonce byte ciphertext byte tag byte plaintext byte associateddata null public void decrypt system readonlyspan nonce system readonlyspan ciphertext system readonlyspan tag system span plaintext system readonlyspan associateddata default system readonlyspan public void dispose public void encrypt byte nonce byte plaintext byte ciphertext byte tag byte associateddata null public void encrypt system readonlyspan nonce system readonlyspan plaintext system span ciphertext system span tag system readonlyspan associateddata default system readonlyspan net type forward csharp matches ref for net public sealed partial class aesgcm system idisposable public aesgcm byte key public aesgcm byte key int tagsizeinbytes public aesgcm system readonlyspan key public aesgcm system readonlyspan key int tagsizeinbytes public static bool issupported get throw null public static system security cryptography keysizes noncebytesizes get throw null public static system security cryptography keysizes tagbytesizes get throw null public int tagsizeinbytes get throw null public void decrypt byte nonce byte ciphertext byte tag byte plaintext byte associateddata null public void decrypt system readonlyspan nonce system readonlyspan ciphertext system readonlyspan tag system span plaintext system readonlyspan associateddata default system readonlyspan public void dispose public void encrypt byte nonce byte plaintext byte ciphertext byte tag byte associateddata null public void encrypt system readonlyspan nonce system readonlyspan plaintext system span ciphertext system span tag system readonlyspan associateddata default system readonlyspan api usage same as net alternative designs this proposal leaves out aesccm this omission is purely due to no known demand for aes ccm in downlevel platforms whereas aes gcm has demonstrated considerable demand we can include aesccm for completeness if desired the ref split is slightly complicated but will improve as net and age out the simplest thing to do is to outbox only the ref from net however i think this is incorrect because it would mean outboxing constructors that are obsolete there is no reason that net framework needs the obsolete constructors we could include both the new and obsolete constructors for net framework if that makes multi targeting easier risks this carries the similar risks of providing any api as a package it carries maintenance burden
| 0
|
714,945
| 24,581,581,756
|
IssuesEvent
|
2022-10-13 16:00:39
|
kubernetes/ingress-nginx
|
https://api.github.com/repos/kubernetes/ingress-nginx
|
closed
|
403 Forbidden Can't pull k8s images in Singapore
|
kind/bug priority/backlog triage/accepted
|
```
0s Warning Failed pod/ingress-nginx-controller-69fbfb4bfd-t24n5 Failed to pull image "registry.k8s.io/ingress-nginx/controller:v1.4.0@sha256:34ee929b111ffc7aa426ffd409af44da48e5a0eea1eb2207994d9e0c0882d143": rpc error: code = Unknown desc = failed to pull and unpack image "registry.k8s.io/ingress-nginx/controller@sha256:34ee929b111ffc7aa426ffd409af44da48e5a0eea1eb2207994d9e0c0882d143": failed to resolve reference "registry.k8s.io/ingress-nginx/controller@sha256:34ee929b111ffc7aa426ffd409af44da48e5a0eea1eb2207994d9e0c0882d143": failed to do request: Head "https://asia-southeast1-docker.pkg.dev/v2/k8s-artifacts-prod/images/ingress-nginx/controller/manifests/sha256:34ee929b111ffc7aa426ffd409af44da48e5a0eea1eb2207994d9e0c0882d143": failed to authorize redirect: failed to fetch anonymous token: unexpected status: 403 Forbidden
```
It works on other regions
|
1.0
|
403 Forbidden Can't pull k8s images in Singapore - ```
0s Warning Failed pod/ingress-nginx-controller-69fbfb4bfd-t24n5 Failed to pull image "registry.k8s.io/ingress-nginx/controller:v1.4.0@sha256:34ee929b111ffc7aa426ffd409af44da48e5a0eea1eb2207994d9e0c0882d143": rpc error: code = Unknown desc = failed to pull and unpack image "registry.k8s.io/ingress-nginx/controller@sha256:34ee929b111ffc7aa426ffd409af44da48e5a0eea1eb2207994d9e0c0882d143": failed to resolve reference "registry.k8s.io/ingress-nginx/controller@sha256:34ee929b111ffc7aa426ffd409af44da48e5a0eea1eb2207994d9e0c0882d143": failed to do request: Head "https://asia-southeast1-docker.pkg.dev/v2/k8s-artifacts-prod/images/ingress-nginx/controller/manifests/sha256:34ee929b111ffc7aa426ffd409af44da48e5a0eea1eb2207994d9e0c0882d143": failed to authorize redirect: failed to fetch anonymous token: unexpected status: 403 Forbidden
```
It works on other regions
|
non_infrastructure
|
forbidden can t pull images in singapore warning failed pod ingress nginx controller failed to pull image registry io ingress nginx controller rpc error code unknown desc failed to pull and unpack image registry io ingress nginx controller failed to resolve reference registry io ingress nginx controller failed to do request head failed to authorize redirect failed to fetch anonymous token unexpected status forbidden it works on other regions
| 0
|
20,251
| 13,778,277,807
|
IssuesEvent
|
2020-10-08 12:15:15
|
gnosis/safe-ios
|
https://api.github.com/repos/gnosis/safe-ios
|
closed
|
Lightweight database migration
|
infrastructure
|
Add the "connected account" key (or different name for it) to the App Settings
# How to test
Please upgrade from the previous app version (prev. app version need to have some data in the app, like some safe loaded).
The app should not crash.
|
1.0
|
Lightweight database migration - Add the "connected account" key (or different name for it) to the App Settings
# How to test
Please upgrade from the previous app version (prev. app version need to have some data in the app, like some safe loaded).
The app should not crash.
|
infrastructure
|
lightweight database migration add the connected account key or different name for it to the app settings how to test please upgrade from the previous app version prev app version need to have some data in the app like some safe loaded the app should not crash
| 1
|
23,504
| 16,343,593,928
|
IssuesEvent
|
2021-05-13 03:09:53
|
dotnet/roslyn-analyzers
|
https://api.github.com/repos/dotnet/roslyn-analyzers
|
closed
|
Packages should use icon, not iconUrl
|
Area-Infrastructure Bug
|
The iconUrl element in the nuspec is deprecated and eventually pushes to nuget.org will stop working. Instead, repos should use icon. This would happen automatically but it looks like roslyn-analyzers manually generates a nuspec rather than using the arcade infra, so it may be a bit more difficult to get working.
|
1.0
|
Packages should use icon, not iconUrl - The iconUrl element in the nuspec is deprecated and eventually pushes to nuget.org will stop working. Instead, repos should use icon. This would happen automatically but it looks like roslyn-analyzers manually generates a nuspec rather than using the arcade infra, so it may be a bit more difficult to get working.
|
infrastructure
|
packages should use icon not iconurl the iconurl element in the nuspec is deprecated and eventually pushes to nuget org will stop working instead repos should use icon this would happen automatically but it looks like roslyn analyzers manually generates a nuspec rather than using the arcade infra so it may be a bit more difficult to get working
| 1
|
28,094
| 22,955,205,358
|
IssuesEvent
|
2022-07-19 10:55:04
|
holaplex/holaplex
|
https://api.github.com/repos/holaplex/holaplex
|
closed
|
Still Using Cards Twitter Link to get Tweets to Show Properly
|
_fix Infrastructure
|
Small detail but I noticed I'm still using the cards twitter link to get the cards to show up on tweets properly.
<img width="594" alt="Screen Shot 2022-05-04 at 10 30 55 AM" src="https://user-images.githubusercontent.com/98030332/166778792-173ce4ef-d72e-4951-91c7-a54d012c6520.png">
Part of the problem feels like its with how the share button on the nft details card is configured but I'm not sure. Def Twitter still not pulling the data before using the cards tool too.
example https://www.holaplex.com/nfts/AcgghCF493ccdXGRfWQxUbX2ZDs4vpcyt6ZepeHJ2AH
|
1.0
|
Still Using Cards Twitter Link to get Tweets to Show Properly - Small detail but I noticed I'm still using the cards twitter link to get the cards to show up on tweets properly.
<img width="594" alt="Screen Shot 2022-05-04 at 10 30 55 AM" src="https://user-images.githubusercontent.com/98030332/166778792-173ce4ef-d72e-4951-91c7-a54d012c6520.png">
Part of the problem feels like its with how the share button on the nft details card is configured but I'm not sure. Def Twitter still not pulling the data before using the cards tool too.
example https://www.holaplex.com/nfts/AcgghCF493ccdXGRfWQxUbX2ZDs4vpcyt6ZepeHJ2AH
|
infrastructure
|
still using cards twitter link to get tweets to show properly small detail but i noticed i m still using the cards twitter link to get the cards to show up on tweets properly img width alt screen shot at am src part of the problem feels like its with how the share button on the nft details card is configured but i m not sure def twitter still not pulling the data before using the cards tool too example
| 1
|
169,084
| 6,394,510,793
|
IssuesEvent
|
2017-08-04 10:29:08
|
OpenEMS/openems
|
https://api.github.com/repos/OpenEMS/openems
|
opened
|
Build all components using maven
|
Component: Backend Component: Edge Component: UI Priority: Medium Type: Enhancement
|
<!--
IF YOU DON'T FILL OUT THE FOLLOWING INFORMATION YOUR ISSUE MIGHT BE CLOSED WITHOUT INVESTIGATING
-->
### Bug Report or Feature Request (mark with an `x`)
```
- [ ] bug report -> please search issues before submitting
- [x] feature request
```
### Bug description or desired functionality.
<!--
What would like to see implemented?
What is the usecase?
-->
All components should be built consistently via maven.
|
1.0
|
Build all components using maven - <!--
IF YOU DON'T FILL OUT THE FOLLOWING INFORMATION YOUR ISSUE MIGHT BE CLOSED WITHOUT INVESTIGATING
-->
### Bug Report or Feature Request (mark with an `x`)
```
- [ ] bug report -> please search issues before submitting
- [x] feature request
```
### Bug description or desired functionality.
<!--
What would like to see implemented?
What is the usecase?
-->
All components should be built consistently via maven.
|
non_infrastructure
|
build all components using maven if you don t fill out the following information your issue might be closed without investigating bug report or feature request mark with an x bug report please search issues before submitting feature request bug description or desired functionality what would like to see implemented what is the usecase all components should be built consistently via maven
| 0
|
205,228
| 7,094,871,808
|
IssuesEvent
|
2018-01-13 09:51:13
|
status-im/status-react
|
https://api.github.com/repos/status-im/status-react
|
closed
|
Password request message is not shown after application is launched after install
|
blocker bug fix them all high-priority release
|
### Description
*Type*: Bug
*Summary*: we have this issue appears from time to time and there are several issues closed for this https://github.com/status-im/status-react/issues/2049, https://github.com/status-im/status-react/issues/1130 for example and this issue is stable to reproduce in `release build ` from Jan 5 2018
https://jenkins.status.im/job/status-react/job/release/job/release%252F0.9.13/1/
#### Expected behavior
I can set up password
#### Actual behavior

### Reproduction
- Open Status
### Additional Information
* Status version: release build 0.9.13 from Jan 5 2018
* Operating System: Android
|
1.0
|
Password request message is not shown after application is launched after install - ### Description
*Type*: Bug
*Summary*: we have this issue appears from time to time and there are several issues closed for this https://github.com/status-im/status-react/issues/2049, https://github.com/status-im/status-react/issues/1130 for example and this issue is stable to reproduce in `release build ` from Jan 5 2018
https://jenkins.status.im/job/status-react/job/release/job/release%252F0.9.13/1/
#### Expected behavior
I can set up password
#### Actual behavior

### Reproduction
- Open Status
### Additional Information
* Status version: release build 0.9.13 from Jan 5 2018
* Operating System: Android
|
non_infrastructure
|
password request message is not shown after application is launched after install description type bug summary we have this issue appears from time to time and there are several issues closed for this for example and this issue is stable to reproduce in release build from jan expected behavior i can set up password actual behavior reproduction open status additional information status version release build from jan operating system android
| 0
|
15,619
| 11,616,876,131
|
IssuesEvent
|
2020-02-26 16:23:05
|
department-of-veterans-affairs/va.gov-cms
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
|
closed
|
Email Sending in BRD & Password Reset Email Functionality
|
Infrastructure โญ๏ธ Infrastructure
|
Reconfigure both systems to use the correct SMTP server
AC:
- CMS uses the correct SMTP server to send email (and email works) in environments deployed by BRD.
-
- ~CMS uses the correct SMTP server to send email (and email works) in PR environments created by CMS-CI.~
***This issue is a migration of VAGOV-526 from Jira to help track 2020 work completed in ZenHub.**
|
2.0
|
Email Sending in BRD & Password Reset Email Functionality - Reconfigure both systems to use the correct SMTP server
AC:
- CMS uses the correct SMTP server to send email (and email works) in environments deployed by BRD.
-
- ~CMS uses the correct SMTP server to send email (and email works) in PR environments created by CMS-CI.~
***This issue is a migration of VAGOV-526 from Jira to help track 2020 work completed in ZenHub.**
|
infrastructure
|
email sending in brd password reset email functionality reconfigure both systems to use the correct smtp server ac cms uses the correct smtp server to send email and email works in environments deployed by brd cms uses the correct smtp server to send email and email works in pr environments created by cms ci this issue is a migration of vagov from jira to help track work completed in zenhub
| 1
|
80,112
| 3,550,865,734
|
IssuesEvent
|
2016-01-20 23:53:41
|
donejs/donejs
|
https://api.github.com/repos/donejs/donejs
|
closed
|
New project references "place-my-order"
|
Priority - High
|
My project is called "scraplog" and somehow my `index.stache` references "place-my-order" in the production `<script>` tag's `main` attribute.
```
<html>
<head>
<title>{{title}}</title>
{{asset "css"}}
</head>
<body>
<can-import from="scraplog/styles.less!" />
<can-import from="scraplog/app" as="viewModel" />
<h1>{{message}}</h1>
{{asset "inline-cache"}}
{{#switch @env.NODE_ENV}}
{{#case "production"}}
<script src="/node_modules/steal/steal.production.js" main="place-my-order/index.stache!done-autorender"></script>
{{/case}}
{{#default}}
<script src="/node_modules/steal/steal.js"></script>
{{/default}}
{{/switch}}
</body>
</html>
```
|
1.0
|
New project references "place-my-order" - My project is called "scraplog" and somehow my `index.stache` references "place-my-order" in the production `<script>` tag's `main` attribute.
```
<html>
<head>
<title>{{title}}</title>
{{asset "css"}}
</head>
<body>
<can-import from="scraplog/styles.less!" />
<can-import from="scraplog/app" as="viewModel" />
<h1>{{message}}</h1>
{{asset "inline-cache"}}
{{#switch @env.NODE_ENV}}
{{#case "production"}}
<script src="/node_modules/steal/steal.production.js" main="place-my-order/index.stache!done-autorender"></script>
{{/case}}
{{#default}}
<script src="/node_modules/steal/steal.js"></script>
{{/default}}
{{/switch}}
</body>
</html>
```
|
non_infrastructure
|
new project references place my order my project is called scraplog and somehow my index stache references place my order in the production tag s main attribute title asset css message asset inline cache switch env node env case production case default default switch
| 0
|
224,529
| 7,471,622,603
|
IssuesEvent
|
2018-04-03 09:51:08
|
caicloud/cyclone
|
https://api.github.com/repos/caicloud/cyclone
|
closed
|
glog version is not correct
|
kind/bug priority/P1
|
**Is this a BUG REPORT or FEATURE REQUEST?**:
/kind bug
**What happened**:
```
ERROR: logging before flag.Parse: I0327 08:44:58.327256 4807 router.go:184] Register pipeline APIs
ERROR: logging before flag.Parse: I0327 08:44:58.327427 4807 router.go:220] Register pipeline record APIs
ERROR: logging before flag.Parse: I0327 08:44:58.327780 4807 router.go:257] Register pipeline records logs APIs
ERROR: logging before flag.Parse: I0327 08:44:58.327945 4807 router.go:287] Register event APIs
ERROR: logging before flag.Parse: I0327 08:44:58.328050 4807 router.go:321] Register cloud APIs
ERROR: logging before flag.Parse: I0327 08:44:58.328188 4807 router.go:346] Register health check API
ERROR: logging before flag.Parse: I0327 08:44:58.328235 4807 router.go:304] Register webhook APIs
```
**What you expected to happen**:
Use the reversion `44145f04b68cf362d9c4df2182967c2275eaefed` of [glog](https://github.com/golang/glog).
**How to reproduce it (as minimally and precisely as possible)**:
**Anything else we need to know?**:
/assign @supereagle
/priority p1
|
1.0
|
glog version is not correct - **Is this a BUG REPORT or FEATURE REQUEST?**:
/kind bug
**What happened**:
```
ERROR: logging before flag.Parse: I0327 08:44:58.327256 4807 router.go:184] Register pipeline APIs
ERROR: logging before flag.Parse: I0327 08:44:58.327427 4807 router.go:220] Register pipeline record APIs
ERROR: logging before flag.Parse: I0327 08:44:58.327780 4807 router.go:257] Register pipeline records logs APIs
ERROR: logging before flag.Parse: I0327 08:44:58.327945 4807 router.go:287] Register event APIs
ERROR: logging before flag.Parse: I0327 08:44:58.328050 4807 router.go:321] Register cloud APIs
ERROR: logging before flag.Parse: I0327 08:44:58.328188 4807 router.go:346] Register health check API
ERROR: logging before flag.Parse: I0327 08:44:58.328235 4807 router.go:304] Register webhook APIs
```
**What you expected to happen**:
Use the reversion `44145f04b68cf362d9c4df2182967c2275eaefed` of [glog](https://github.com/golang/glog).
**How to reproduce it (as minimally and precisely as possible)**:
**Anything else we need to know?**:
/assign @supereagle
/priority p1
|
non_infrastructure
|
glog version is not correct is this a bug report or feature request kind bug what happened error logging before flag parse router go register pipeline apis error logging before flag parse router go register pipeline record apis error logging before flag parse router go register pipeline records logs apis error logging before flag parse router go register event apis error logging before flag parse router go register cloud apis error logging before flag parse router go register health check api error logging before flag parse router go register webhook apis what you expected to happen use the reversion of how to reproduce it as minimally and precisely as possible anything else we need to know assign supereagle priority
| 0
|
24,390
| 17,187,089,706
|
IssuesEvent
|
2021-07-16 04:50:45
|
source-academy/frontend
|
https://api.github.com/repos/source-academy/frontend
|
closed
|
Refactor evalCode saga for Autograder run-all functionality
|
_infrastructure _postponed _refactor
|
## Refactor evalCode saga for Autograder run-all functionality
Summary: Deferred refactoring to continue support for Mission Autograder run-all functionality from #801.
### Todos
- Remove `sideContentActiveTab` attribute from Redux store along with the associated action `UPDATE_ACTIVE_TAB` with its action creator and reducer, in line with the project to clean up the Redux state tree
- Remove conditional behaviour in `evalCode` saga for run-all testcase functionality
- Re-implement run-all testcase functionality in line with #607 by constructing a different 'Run' button in assessment and grading workspaces, to be passed to `ControlBar`
- This button will contain its own internal `activeTab` prop inherited from the state of the containing workspace, and will invoke either `EVAL_EDITOR` or `EVAL_EDITOR_WITH_TESTCASES` depending on the value of its `activeTab` prop
- `activeTab` state of the containing workspace will be updated by the `onChange` handler of the respective workspace's `SideContent` component
- Add a new action `EVAL_EDITOR_WITH_TESTCASES` with its action creator, reducer and saga to be invoked by the alternative 'Run' button
- Repair all broken tests from the refactor
|
1.0
|
Refactor evalCode saga for Autograder run-all functionality - ## Refactor evalCode saga for Autograder run-all functionality
Summary: Deferred refactoring to continue support for Mission Autograder run-all functionality from #801.
### Todos
- Remove `sideContentActiveTab` attribute from Redux store along with the associated action `UPDATE_ACTIVE_TAB` with its action creator and reducer, in line with the project to clean up the Redux state tree
- Remove conditional behaviour in `evalCode` saga for run-all testcase functionality
- Re-implement run-all testcase functionality in line with #607 by constructing a different 'Run' button in assessment and grading workspaces, to be passed to `ControlBar`
- This button will contain its own internal `activeTab` prop inherited from the state of the containing workspace, and will invoke either `EVAL_EDITOR` or `EVAL_EDITOR_WITH_TESTCASES` depending on the value of its `activeTab` prop
- `activeTab` state of the containing workspace will be updated by the `onChange` handler of the respective workspace's `SideContent` component
- Add a new action `EVAL_EDITOR_WITH_TESTCASES` with its action creator, reducer and saga to be invoked by the alternative 'Run' button
- Repair all broken tests from the refactor
|
infrastructure
|
refactor evalcode saga for autograder run all functionality refactor evalcode saga for autograder run all functionality summary deferred refactoring to continue support for mission autograder run all functionality from todos remove sidecontentactivetab attribute from redux store along with the associated action update active tab with its action creator and reducer in line with the project to clean up the redux state tree remove conditional behaviour in evalcode saga for run all testcase functionality re implement run all testcase functionality in line with by constructing a different run button in assessment and grading workspaces to be passed to controlbar this button will contain its own internal activetab prop inherited from the state of the containing workspace and will invoke either eval editor or eval editor with testcases depending on the value of its activetab prop activetab state of the containing workspace will be updated by the onchange handler of the respective workspace s sidecontent component add a new action eval editor with testcases with its action creator reducer and saga to be invoked by the alternative run button repair all broken tests from the refactor
| 1
|
1,811
| 3,479,203,614
|
IssuesEvent
|
2015-12-28 18:29:51
|
codynguyen1116/brickimedia-design-guide
|
https://api.github.com/repos/codynguyen1116/brickimedia-design-guide
|
reopened
|
Alternative to LESS?
|
discussion enhancement security
|
I've never used LESS before but from what I can see of it here, I'm not impressed nor satisfied with it. First of all, and most noticeably, even something as simple as ```index.html```'s ```#header``` isn't getting any styles anymore, which it was yesterday. Next, it does not fail gracefully with XHR, which it shouldn't even need to be doing. It's looking in the wrong places for files, and I've tried using local sources (which shouldn't even be loaded over XHR at all...) and absolute URLs, neither of which are working. Even with an <b>absolute</b> URL, it's still trying to find resources somewhere else. Wtf??

I can't even debug this anymore because it's basically the worst library I've ever used, so if there are any alternatives that don't involve insecure loading of local resources and utter retardedness, that'd be great to switch to.
|
True
|
Alternative to LESS? - I've never used LESS before but from what I can see of it here, I'm not impressed nor satisfied with it. First of all, and most noticeably, even something as simple as ```index.html```'s ```#header``` isn't getting any styles anymore, which it was yesterday. Next, it does not fail gracefully with XHR, which it shouldn't even need to be doing. It's looking in the wrong places for files, and I've tried using local sources (which shouldn't even be loaded over XHR at all...) and absolute URLs, neither of which are working. Even with an <b>absolute</b> URL, it's still trying to find resources somewhere else. Wtf??

I can't even debug this anymore because it's basically the worst library I've ever used, so if there are any alternatives that don't involve insecure loading of local resources and utter retardedness, that'd be great to switch to.
|
non_infrastructure
|
alternative to less i ve never used less before but from what i can see of it here i m not impressed nor satisfied with it first of all and most noticeably even something as simple as index html s header isn t getting any styles anymore which it was yesterday next it does not fail gracefully with xhr which it shouldn t even need to be doing it s looking in the wrong places for files and i ve tried using local sources which shouldn t even be loaded over xhr at all and absolute urls neither of which are working even with an absolute url it s still trying to find resources somewhere else wtf i can t even debug this anymore because it s basically the worst library i ve ever used so if there are any alternatives that don t involve insecure loading of local resources and utter retardedness that d be great to switch to
| 0
|
30,235
| 24,672,230,569
|
IssuesEvent
|
2022-10-18 14:30:13
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
closed
|
dotnet pack does not respect target framework conditions
|
area-Infrastructure-libraries untriaged
|
### Description
I have multitargeted csproj and I want to build and pack only the single target but `dotnet pack` always ignores passed configuration and packs all targets.
### Reproduction Steps
```
<Project Sdk="Microsoft.NET.Sdk">
<PropertyGroup>
<TargetFrameworks>net6.0;netstandard2.0</TargetFrameworks>
<Configurations>Net6;Netstandard20</Configurations>
</PropertyGroup>
<PropertyGroup Condition="'$(Configuration)' == 'Net6'">
<TargetFramework>net6.0</TargetFramework>
</PropertyGroup>
<PropertyGroup Condition="'$(Configuration)' == 'Netstandard20'">
<TargetFramework>netstandard2.0</TargetFramework>
</PropertyGroup>
</Project>
```
The following command
`dotnet pack -c Netstandard20`
results in error
`C:\Program Files\dotnet\sdk\6.0.201\Sdks\NuGet.Build.Tasks.Pack\build\NuGet.Build.Tasks.Pack.targets(221,5): error NU5026: The file 'E:\cm_1\drive\TestApp\bin\Netstandard20\net6.0\TestApp.dll' to be packed was not found on disk. [E:\cm_1\drive\TestApp\TestApp.csproj]`
because net6.0 configuration was not built and folder net6.0 is empty indeed.
I also tried to publish before packing (without build), but dotnet pack still does not seem to respect passed configuration.
1. `dotnet publish -c Netstandard20` - this command builds requested configuration and publishes it to `bin\Netstandard20\netstandard2.0\publish`.
2. `dotnet pack --no-build /p:OutputPath=bin\Netstandard20\netstandard2.0\publish\` - and this command produces nuget package which containts both folders `lib\net6.0` and `lib\netstandard2.0` with same published binaries which also looks incorrect since I wanted to publish and pack only netstandard2.0 configuration
### Expected behavior
`dotnet pack -c Netstandard20` to pack only the requested configuration and not try to do anything with the net6.0 target.
### Actual behavior
`dotnet pack -c Netstandard20` builds and packs all configurations listed in TargetFrameworks even though the following csproj conditions constrain the build only to the single target for passed configuration.
### Regression?
_No response_
### Known Workarounds
Can't figure out any, but would be glad if any would be suggested.
### Configuration
_No response_
### Other information
_No response_
|
1.0
|
dotnet pack does not respect target framework conditions - ### Description
I have multitargeted csproj and I want to build and pack only the single target but `dotnet pack` always ignores passed configuration and packs all targets.
### Reproduction Steps
```
<Project Sdk="Microsoft.NET.Sdk">
<PropertyGroup>
<TargetFrameworks>net6.0;netstandard2.0</TargetFrameworks>
<Configurations>Net6;Netstandard20</Configurations>
</PropertyGroup>
<PropertyGroup Condition="'$(Configuration)' == 'Net6'">
<TargetFramework>net6.0</TargetFramework>
</PropertyGroup>
<PropertyGroup Condition="'$(Configuration)' == 'Netstandard20'">
<TargetFramework>netstandard2.0</TargetFramework>
</PropertyGroup>
</Project>
```
The following command
`dotnet pack -c Netstandard20`
results in error
`C:\Program Files\dotnet\sdk\6.0.201\Sdks\NuGet.Build.Tasks.Pack\build\NuGet.Build.Tasks.Pack.targets(221,5): error NU5026: The file 'E:\cm_1\drive\TestApp\bin\Netstandard20\net6.0\TestApp.dll' to be packed was not found on disk. [E:\cm_1\drive\TestApp\TestApp.csproj]`
because net6.0 configuration was not built and folder net6.0 is empty indeed.
I also tried to publish before packing (without build), but dotnet pack still does not seem to respect passed configuration.
1. `dotnet publish -c Netstandard20` - this command builds requested configuration and publishes it to `bin\Netstandard20\netstandard2.0\publish`.
2. `dotnet pack --no-build /p:OutputPath=bin\Netstandard20\netstandard2.0\publish\` - and this command produces nuget package which containts both folders `lib\net6.0` and `lib\netstandard2.0` with same published binaries which also looks incorrect since I wanted to publish and pack only netstandard2.0 configuration
### Expected behavior
`dotnet pack -c Netstandard20` to pack only the requested configuration and not try to do anything with the net6.0 target.
### Actual behavior
`dotnet pack -c Netstandard20` builds and packs all configurations listed in TargetFrameworks even though the following csproj conditions constrain the build only to the single target for passed configuration.
### Regression?
_No response_
### Known Workarounds
Can't figure out any, but would be glad if any would be suggested.
### Configuration
_No response_
### Other information
_No response_
|
infrastructure
|
dotnet pack does not respect target framework conditions description i have multitargeted csproj and i want to build and pack only the single target but dotnet pack always ignores passed configuration and packs all targets reproduction steps the following command dotnet pack c results in error c program files dotnet sdk sdks nuget build tasks pack build nuget build tasks pack targets error the file e cm drive testapp bin testapp dll to be packed was not found on disk because configuration was not built and folder is empty indeed i also tried to publish before packing without build but dotnet pack still does not seem to respect passed configuration dotnet publish c this command builds requested configuration and publishes it to bin publish dotnet pack no build p outputpath bin publish and this command produces nuget package which containts both folders lib and lib with same published binaries which also looks incorrect since i wanted to publish and pack only configuration expected behavior dotnet pack c to pack only the requested configuration and not try to do anything with the target actual behavior dotnet pack c builds and packs all configurations listed in targetframeworks even though the following csproj conditions constrain the build only to the single target for passed configuration regression no response known workarounds can t figure out any but would be glad if any would be suggested configuration no response other information no response
| 1
|
26,032
| 19,580,873,713
|
IssuesEvent
|
2022-01-04 21:07:44
|
astropy/astropy-tutorials
|
https://api.github.com/repos/astropy/astropy-tutorials
|
opened
|
Add a section navigation sidebar to the tutorial HTML page template
|
infrastructure
|
Let's add a sidebar to our HTML page template (see also #513) that includes links to the section titles. This would need some code that parses the markdown content and build the section hierarchy.
Doing this will help a lot with longer tutorials by providing an automatic overview of the content and an affordance for moving around quickly within the tutorial.
|
1.0
|
Add a section navigation sidebar to the tutorial HTML page template - Let's add a sidebar to our HTML page template (see also #513) that includes links to the section titles. This would need some code that parses the markdown content and build the section hierarchy.
Doing this will help a lot with longer tutorials by providing an automatic overview of the content and an affordance for moving around quickly within the tutorial.
|
infrastructure
|
add a section navigation sidebar to the tutorial html page template let s add a sidebar to our html page template see also that includes links to the section titles this would need some code that parses the markdown content and build the section hierarchy doing this will help a lot with longer tutorials by providing an automatic overview of the content and an affordance for moving around quickly within the tutorial
| 1
|
28,661
| 23,426,987,808
|
IssuesEvent
|
2022-08-14 14:37:15
|
ThorstenSauter/NoPlan
|
https://api.github.com/repos/ThorstenSauter/NoPlan
|
closed
|
Bring role assignments for RBAC under management by Pulumi
|
infrastructure
|
## `AAD` objects
- `CI/CD` service principal - the service principal running in GitHub Actions
- `Developer group` - the `AAD group` containing the application developers
- `User assigned managed identity` - the identity of resources that need to access other resources
## Resources and their role assignments:
| Resource | Role | AAD object |
| :----------------------------- | :-------------------------------- | :------------------------------ |
| **Azure Container Registry** | `AcrPull` | User assigned managed identity |
| | `AcrPush` | `CI/CD` service principal |
| **Azure App Configuration** | `App Configuration Data Owner` | Developer group |
| | `App Configuration Data Reader` | User assigned managed identity |
| Application **Key Vault** | `Key Vault Administrator` | Developer group |
| | `Key Vault Secrets User` | User assigned managed identity |
| Integration test **Key Vault** | `Key Vault Administrator` | Developer group |
| | `Key Vault Secrets User` | `CI/CD` service principal |
| **Azure Service Bus** | `Azure Service Bus Data Owner` | Developer group |
| | `Azure Service Bus Data Sender` | User assigned managed identity |
| | `Azure Service Bus Data Receiver` | User assigned managed identity |
|
1.0
|
Bring role assignments for RBAC under management by Pulumi - ## `AAD` objects
- `CI/CD` service principal - the service principal running in GitHub Actions
- `Developer group` - the `AAD group` containing the application developers
- `User assigned managed identity` - the identity of resources that need to access other resources
## Resources and their role assignments:
| Resource | Role | AAD object |
| :----------------------------- | :-------------------------------- | :------------------------------ |
| **Azure Container Registry** | `AcrPull` | User assigned managed identity |
| | `AcrPush` | `CI/CD` service principal |
| **Azure App Configuration** | `App Configuration Data Owner` | Developer group |
| | `App Configuration Data Reader` | User assigned managed identity |
| Application **Key Vault** | `Key Vault Administrator` | Developer group |
| | `Key Vault Secrets User` | User assigned managed identity |
| Integration test **Key Vault** | `Key Vault Administrator` | Developer group |
| | `Key Vault Secrets User` | `CI/CD` service principal |
| **Azure Service Bus** | `Azure Service Bus Data Owner` | Developer group |
| | `Azure Service Bus Data Sender` | User assigned managed identity |
| | `Azure Service Bus Data Receiver` | User assigned managed identity |
|
infrastructure
|
bring role assignments for rbac under management by pulumi aad objects ci cd service principal the service principal running in github actions developer group the aad group containing the application developers user assigned managed identity the identity of resources that need to access other resources resources and their role assignments resource role aad object azure container registry acrpull user assigned managed identity acrpush ci cd service principal azure app configuration app configuration data owner developer group app configuration data reader user assigned managed identity application key vault key vault administrator developer group key vault secrets user user assigned managed identity integration test key vault key vault administrator developer group key vault secrets user ci cd service principal azure service bus azure service bus data owner developer group azure service bus data sender user assigned managed identity azure service bus data receiver user assigned managed identity
| 1
|
28,767
| 12,963,098,461
|
IssuesEvent
|
2020-07-20 18:13:34
|
microsoft/BotFramework-WebChat
|
https://api.github.com/repos/microsoft/BotFramework-WebChat
|
closed
|
DirectLine domain one endpoint websocket inside - standard endpoint
|
Bot Services customer-replied-to customer-reported
|
Hi
Original ticket was closed, but I sill have questions ...
_Originally posted by @vglu in https://github.com/microsoft/BotFramework-WebChat/issues/3188#issuecomment-638112705_
Direct Line App service Extension.
First of all - thank you for your advise - I really appreciate it.
Let me detail explain my plan.
I have got standard
1 QnA service
2 Published bot from QnA service (without any changes)
3 One intranet page call DirectLine with `domain = 'myDomain/v3/directline`
4 One nginx proxy server with rules to redirect myDomain to Azure bot service
Just only one my local page with changes, all others from the box. This mean, upgrade for new version, bug fixing, support all this services I have got from MS.
Now, if I'll use `Direct Line App service Extension`
Yours suggestion
If I'll use it - I should place on-prem one more server. In case with proxy - I have got only proxy server and all others (support, traffic, and so on) I have got without changes from Azure. In my case it is - QnA + publish bot (without any changes) and one page for users. In your suggestion I should modify bot and more more ....
Additional not too geed describe how to setup this functionality from begin to end. May be you have additional resources where I can read and try your solution?
Lets return back to question ....
If I pass my domain to DirectLine class - why I have got connection to external socket in response?
Thanks
Regards,
Vitaliy
_Originally posted by @vglu in https://github.com/microsoft/BotFramework-WebChat/issues/3188#issuecomment-638112705_
|
1.0
|
DirectLine domain one endpoint websocket inside - standard endpoint - Hi
Original ticket was closed, but I sill have questions ...
_Originally posted by @vglu in https://github.com/microsoft/BotFramework-WebChat/issues/3188#issuecomment-638112705_
Direct Line App service Extension.
First of all - thank you for your advise - I really appreciate it.
Let me detail explain my plan.
I have got standard
1 QnA service
2 Published bot from QnA service (without any changes)
3 One intranet page call DirectLine with `domain = 'myDomain/v3/directline`
4 One nginx proxy server with rules to redirect myDomain to Azure bot service
Just only one my local page with changes, all others from the box. This mean, upgrade for new version, bug fixing, support all this services I have got from MS.
Now, if I'll use `Direct Line App service Extension`
Yours suggestion
If I'll use it - I should place on-prem one more server. In case with proxy - I have got only proxy server and all others (support, traffic, and so on) I have got without changes from Azure. In my case it is - QnA + publish bot (without any changes) and one page for users. In your suggestion I should modify bot and more more ....
Additional not too geed describe how to setup this functionality from begin to end. May be you have additional resources where I can read and try your solution?
Lets return back to question ....
If I pass my domain to DirectLine class - why I have got connection to external socket in response?
Thanks
Regards,
Vitaliy
_Originally posted by @vglu in https://github.com/microsoft/BotFramework-WebChat/issues/3188#issuecomment-638112705_
|
non_infrastructure
|
directline domain one endpoint websocket inside standard endpoint hi original ticket was closed but i sill have questions originally posted by vglu in direct line app service extension first of all thank you for your advise i really appreciate it let me detail explain my plan i have got standard qna service published bot from qna service without any changes one intranet page call directline with domain mydomain directline one nginx proxy server with rules to redirect mydomain to azure bot service just only one my local page with changes all others from the box this mean upgrade for new version bug fixing support all this services i have got from ms now if i ll use direct line app service extension yours suggestion if i ll use it i should place on prem one more server in case with proxy i have got only proxy server and all others support traffic and so on i have got without changes from azure in my case it is qna publish bot without any changes and one page for users in your suggestion i should modify bot and more more additional not too geed describe how to setup this functionality from begin to end may be you have additional resources where i can read and try your solution lets return back to question if i pass my domain to directline class why i have got connection to external socket in response thanks regards vitaliy originally posted by vglu in
| 0
|
35,515
| 12,344,613,403
|
IssuesEvent
|
2020-05-15 07:22:51
|
ondrejrozsypal/test
|
https://api.github.com/repos/ondrejrozsypal/test
|
closed
|
CVE-2016-9014 (High) detected in Django-1.8.15-py2.py3-none-any.whl
|
security vulnerability
|
## CVE-2016-9014 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Django-1.8.15-py2.py3-none-any.whl</b></p></summary>
<p>A high-level Python Web framework that encourages rapid development and clean, pragmatic design.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/f3/bb/865626b78c2e87acd4dc1bd8f8cb0a49582ec65e40968f3d24f594bba600/Django-1.8.15-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/f3/bb/865626b78c2e87acd4dc1bd8f8cb0a49582ec65e40968f3d24f594bba600/Django-1.8.15-py2.py3-none-any.whl</a></p>
<p>Path to dependency file: /tmp/ws-scm/test/pip-requirements.txt</p>
<p>Path to vulnerable library: /test/pip-requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **Django-1.8.15-py2.py3-none-any.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ondrejrozsypal/test/commit/cd4b66cdcc157d793ca27f34180fe710ff3fddd8">cd4b66cdcc157d793ca27f34180fe710ff3fddd8</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Django before 1.8.x before 1.8.16, 1.9.x before 1.9.11, and 1.10.x before 1.10.3, when settings.DEBUG is True, allow remote attackers to conduct DNS rebinding attacks by leveraging failure to validate the HTTP Host header against settings.ALLOWED_HOSTS.
<p>Publish Date: 2016-12-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-9014>CVE-2016-9014</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-9014">https://nvd.nist.gov/vuln/detail/CVE-2016-9014</a></p>
<p>Release Date: 2016-12-09</p>
<p>Fix Resolution: 1.8.16,1.9.11,1.10.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2016-9014 (High) detected in Django-1.8.15-py2.py3-none-any.whl - ## CVE-2016-9014 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Django-1.8.15-py2.py3-none-any.whl</b></p></summary>
<p>A high-level Python Web framework that encourages rapid development and clean, pragmatic design.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/f3/bb/865626b78c2e87acd4dc1bd8f8cb0a49582ec65e40968f3d24f594bba600/Django-1.8.15-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/f3/bb/865626b78c2e87acd4dc1bd8f8cb0a49582ec65e40968f3d24f594bba600/Django-1.8.15-py2.py3-none-any.whl</a></p>
<p>Path to dependency file: /tmp/ws-scm/test/pip-requirements.txt</p>
<p>Path to vulnerable library: /test/pip-requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **Django-1.8.15-py2.py3-none-any.whl** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ondrejrozsypal/test/commit/cd4b66cdcc157d793ca27f34180fe710ff3fddd8">cd4b66cdcc157d793ca27f34180fe710ff3fddd8</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Django before 1.8.x before 1.8.16, 1.9.x before 1.9.11, and 1.10.x before 1.10.3, when settings.DEBUG is True, allow remote attackers to conduct DNS rebinding attacks by leveraging failure to validate the HTTP Host header against settings.ALLOWED_HOSTS.
<p>Publish Date: 2016-12-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-9014>CVE-2016-9014</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-9014">https://nvd.nist.gov/vuln/detail/CVE-2016-9014</a></p>
<p>Release Date: 2016-12-09</p>
<p>Fix Resolution: 1.8.16,1.9.11,1.10.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_infrastructure
|
cve high detected in django none any whl cve high severity vulnerability vulnerable library django none any whl a high level python web framework that encourages rapid development and clean pragmatic design library home page a href path to dependency file tmp ws scm test pip requirements txt path to vulnerable library test pip requirements txt dependency hierarchy x django none any whl vulnerable library found in head commit a href vulnerability details django before x before x before and x before when settings debug is true allow remote attackers to conduct dns rebinding attacks by leveraging failure to validate the http host header against settings allowed hosts publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
212,884
| 23,956,918,061
|
IssuesEvent
|
2022-09-12 15:37:47
|
mendts-workshop1/WrzocHoo
|
https://api.github.com/repos/mendts-workshop1/WrzocHoo
|
closed
|
mysql-connector-java-8.0.28.jar: 1 vulnerabilities (highest severity is: 5.5) - autoclosed
|
security vulnerability
|
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mysql-connector-java-8.0.28.jar</b></p></summary>
<p></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/google/protobuf/protobuf-java/3.11.4/protobuf-java-3.11.4.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/mendts-workshop1/WrzocHoo/commit/ce600927af47fb74fe3900cc32059066d77c06aa">ce600927af47fb74fe3900cc32059066d77c06aa</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2021-22569](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22569) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.5 | protobuf-java-3.11.4.jar | Transitive | 8.0.29 | ✅ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2021-22569</summary>
### Vulnerable Library - <b>protobuf-java-3.11.4.jar</b></p>
<p>Core Protocol Buffers library. Protocol Buffers are a way of encoding structured data in an
efficient yet extensible format.</p>
<p>Library home page: <a href="https://developers.google.com/protocol-buffers/">https://developers.google.com/protocol-buffers/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/google/protobuf/protobuf-java/3.11.4/protobuf-java-3.11.4.jar</p>
<p>
Dependency Hierarchy:
- mysql-connector-java-8.0.28.jar (Root Library)
- :x: **protobuf-java-3.11.4.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/mendts-workshop1/WrzocHoo/commit/ce600927af47fb74fe3900cc32059066d77c06aa">ce600927af47fb74fe3900cc32059066d77c06aa</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
An issue in protobuf-java allowed the interleaving of com.google.protobuf.UnknownFieldSet fields in such a way that would be processed out of order. A small malicious payload can occupy the parser for several minutes by creating large numbers of short-lived objects that cause frequent, repeated pauses. We recommend upgrading libraries beyond the vulnerable versions.
<p>Publish Date: 2022-01-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22569>CVE-2021-22569</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-wrvw-hg22-4m67">https://github.com/advisories/GHSA-wrvw-hg22-4m67</a></p>
<p>Release Date: 2022-01-10</p>
<p>Fix Resolution (com.google.protobuf:protobuf-java): 3.16.1</p>
<p>Direct dependency fix Resolution (mysql:mysql-connector-java): 8.0.29</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p>
|
True
|
mysql-connector-java-8.0.28.jar: 1 vulnerabilities (highest severity is: 5.5) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mysql-connector-java-8.0.28.jar</b></p></summary>
<p></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/google/protobuf/protobuf-java/3.11.4/protobuf-java-3.11.4.jar</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/mendts-workshop1/WrzocHoo/commit/ce600927af47fb74fe3900cc32059066d77c06aa">ce600927af47fb74fe3900cc32059066d77c06aa</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2021-22569](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22569) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.5 | protobuf-java-3.11.4.jar | Transitive | 8.0.29 | ✅ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2021-22569</summary>
### Vulnerable Library - <b>protobuf-java-3.11.4.jar</b></p>
<p>Core Protocol Buffers library. Protocol Buffers are a way of encoding structured data in an
efficient yet extensible format.</p>
<p>Library home page: <a href="https://developers.google.com/protocol-buffers/">https://developers.google.com/protocol-buffers/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/google/protobuf/protobuf-java/3.11.4/protobuf-java-3.11.4.jar</p>
<p>
Dependency Hierarchy:
- mysql-connector-java-8.0.28.jar (Root Library)
- :x: **protobuf-java-3.11.4.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/mendts-workshop1/WrzocHoo/commit/ce600927af47fb74fe3900cc32059066d77c06aa">ce600927af47fb74fe3900cc32059066d77c06aa</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
An issue in protobuf-java allowed the interleaving of com.google.protobuf.UnknownFieldSet fields in such a way that would be processed out of order. A small malicious payload can occupy the parser for several minutes by creating large numbers of short-lived objects that cause frequent, repeated pauses. We recommend upgrading libraries beyond the vulnerable versions.
<p>Publish Date: 2022-01-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22569>CVE-2021-22569</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-wrvw-hg22-4m67">https://github.com/advisories/GHSA-wrvw-hg22-4m67</a></p>
<p>Release Date: 2022-01-10</p>
<p>Fix Resolution (com.google.protobuf:protobuf-java): 3.16.1</p>
<p>Direct dependency fix Resolution (mysql:mysql-connector-java): 8.0.29</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p>
|
non_infrastructure
|
mysql connector java jar vulnerabilities highest severity is autoclosed vulnerable library mysql connector java jar path to dependency file pom xml path to vulnerable library home wss scanner repository com google protobuf protobuf java protobuf java jar found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available medium protobuf java jar transitive details cve vulnerable library protobuf java jar core protocol buffers library protocol buffers are a way of encoding structured data in an efficient yet extensible format library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository com google protobuf protobuf java protobuf java jar dependency hierarchy mysql connector java jar root library x protobuf java jar vulnerable library found in head commit a href found in base branch master vulnerability details an issue in protobuf java allowed the interleaving of com google protobuf unknownfieldset fields in such a way that would be processed out of order a small malicious payload can occupy the parser for several minutes by creating large numbers of short lived objects that cause frequent repeated pauses we recommend upgrading libraries beyond the vulnerable versions publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com google protobuf protobuf java direct dependency fix resolution mysql mysql connector java rescue worker helmet automatic remediation is available for this issue rescue worker helmet automatic remediation is available for this issue
| 0
|
13,160
| 10,131,822,337
|
IssuesEvent
|
2019-08-01 20:33:55
|
HumanCellAtlas/secondary-analysis
|
https://api.github.com/repos/HumanCellAtlas/secondary-analysis
|
reopened
|
Add maxRetries to HCA workflow options
|
infrastructure
|
The Optimus workflow occasionally fails due to transient errors that can be fixed with automated retries. A default value for the maxRetries runtime parameter can be configured in the workflow options file.
Have Lira set a default value based on what is specified in its config, so that it is automatically applied to every task. This also requires explicitly setting the maxRetries value to zero in submit wdl tasks that should not be retried.
โIssue is synchronized with this [Jira Story](https://broadinstitute.atlassian.net/browse/GH-282)
|
1.0
|
Add maxRetries to HCA workflow options - The Optimus workflow occasionally fails due to transient errors that can be fixed with automated retries. A default value for the maxRetries runtime parameter can be configured in the workflow options file.
Have Lira set a default value based on what is specified in its config, so that it is automatically applied to every task. This also requires explicitly setting the maxRetries value to zero in submit wdl tasks that should not be retried.
โIssue is synchronized with this [Jira Story](https://broadinstitute.atlassian.net/browse/GH-282)
|
infrastructure
|
add maxretries to hca workflow options the optimus workflow occasionally fails due to transient errors that can be fixed with automated retries a default value for the maxretries runtime parameter can be configured in the workflow options file have lira set a default value based on what is specified in its config so that it is automatically applied to every task this also requires explicitly setting the maxretries value to zero in submit wdl tasks that should not be retried โissue is synchronized with this
| 1
|
125,049
| 12,246,027,769
|
IssuesEvent
|
2020-05-05 13:52:38
|
BDDAvanzadas-P-AA2020/casoalquilervideos-davidalejo
|
https://api.github.com/repos/BDDAvanzadas-P-AA2020/casoalquilervideos-davidalejo
|
opened
|
Retroalimentaciรณn del caso alquiler de videos
|
documentation
|
Estimado David estรก muy bien el anรกlisis de transacciones. Lo que no veo es la distribuciรณn de almacenamiento considerando que tenemos un array de discos.
Saludos cordiales
|
1.0
|
Retroalimentaciรณn del caso alquiler de videos - Estimado David estรก muy bien el anรกlisis de transacciones. Lo que no veo es la distribuciรณn de almacenamiento considerando que tenemos un array de discos.
Saludos cordiales
|
non_infrastructure
|
retroalimentaciรณn del caso alquiler de videos estimado david estรก muy bien el anรกlisis de transacciones lo que no veo es la distribuciรณn de almacenamiento considerando que tenemos un array de discos saludos cordiales
| 0
|
17,886
| 12,682,135,609
|
IssuesEvent
|
2020-06-19 16:45:27
|
cmu-db/terrier
|
https://api.github.com/repos/cmu-db/terrier
|
closed
|
Convert the nightly build into a Jenkinsfile
|
infrastructure question/discussion
|
The nightly build is currently just a script written on Jenkins. The MSE Team wants to convert it into a file stored in the repo so that we could have `Stage Step` in our nightly build that runs OLTPbench and other jobs in the future.
The way to do that is to add another `Jenkinsfile` (with a different name, maybe `Jenkinsfile-nightly`?) And configure the [terrier-nightly pipeline](http://jenkins.db.cs.cmu.edu:8080/blue/organizations/jenkins/terrier-nightly/activity). In the `Jenkins Configure - Build Configuration`, we change the Script Path:

The MSE Team is working on the nightly Jenkinsfile. We need @crd477 and @apavlo 's review after we finish it and update the Jenkins Pipeline configuration so that we could test it.
Feel free to discuss this.
|
1.0
|
Convert the nightly build into a Jenkinsfile - The nightly build is currently just a script written on Jenkins. The MSE Team wants to convert it into a file stored in the repo so that we could have `Stage Step` in our nightly build that runs OLTPbench and other jobs in the future.
The way to do that is to add another `Jenkinsfile` (with a different name, maybe `Jenkinsfile-nightly`?) And configure the [terrier-nightly pipeline](http://jenkins.db.cs.cmu.edu:8080/blue/organizations/jenkins/terrier-nightly/activity). In the `Jenkins Configure - Build Configuration`, we change the Script Path:

The MSE Team is working on the nightly Jenkinsfile. We need @crd477 and @apavlo 's review after we finish it and update the Jenkins Pipeline configuration so that we could test it.
Feel free to discuss this.
|
infrastructure
|
convert the nightly build into a jenkinsfile the nightly build is currently just a script written on jenkins the mse team wants to convert it into a file stored in the repo so that we could have stage step in our nightly build that runs oltpbench and other jobs in the future the way to do that is to add another jenkinsfile with a different name maybe jenkinsfile nightly and configure the in the jenkins configure build configuration we change the script path the mse team is working on the nightly jenkinsfile we need and apavlo s review after we finish it and update the jenkins pipeline configuration so that we could test it feel free to discuss this
| 1
|
231,662
| 25,525,487,797
|
IssuesEvent
|
2022-11-29 01:38:22
|
kapseliboi/WeiPay
|
https://api.github.com/repos/kapseliboi/WeiPay
|
reopened
|
CVE-2022-37598 (High) detected in uglify-js-3.4.9.tgz
|
security vulnerability
|
## CVE-2022-37598 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>uglify-js-3.4.9.tgz</b></p></summary>
<p>JavaScript parser, mangler/compressor and beautifier toolkit</p>
<p>Library home page: <a href="https://registry.npmjs.org/uglify-js/-/uglify-js-3.4.9.tgz">https://registry.npmjs.org/uglify-js/-/uglify-js-3.4.9.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/handlebars/node_modules/uglify-js/package.json</p>
<p>
Dependency Hierarchy:
- jest-21.2.1.tgz (Root Library)
- jest-cli-21.2.1.tgz
- istanbul-api-1.3.7.tgz
- istanbul-reports-1.5.1.tgz
- handlebars-4.1.0.tgz
- :x: **uglify-js-3.4.9.tgz** (Vulnerable Library)
<p>Found in base branch: <b>stable</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
** DISPUTED ** Prototype pollution vulnerability in function DEFNODE in ast.js in mishoo UglifyJS 3.13.2 via the name variable in ast.js. NOTE: the vendor considers this an invalid report.
<p>Publish Date: 2022-10-20
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-37598>CVE-2022-37598</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-10-20</p>
<p>Fix Resolution (uglify-js): 3.13.10</p>
<p>Direct dependency fix Resolution (jest): 22.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2022-37598 (High) detected in uglify-js-3.4.9.tgz - ## CVE-2022-37598 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>uglify-js-3.4.9.tgz</b></p></summary>
<p>JavaScript parser, mangler/compressor and beautifier toolkit</p>
<p>Library home page: <a href="https://registry.npmjs.org/uglify-js/-/uglify-js-3.4.9.tgz">https://registry.npmjs.org/uglify-js/-/uglify-js-3.4.9.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/handlebars/node_modules/uglify-js/package.json</p>
<p>
Dependency Hierarchy:
- jest-21.2.1.tgz (Root Library)
- jest-cli-21.2.1.tgz
- istanbul-api-1.3.7.tgz
- istanbul-reports-1.5.1.tgz
- handlebars-4.1.0.tgz
- :x: **uglify-js-3.4.9.tgz** (Vulnerable Library)
<p>Found in base branch: <b>stable</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
** DISPUTED ** Prototype pollution vulnerability in function DEFNODE in ast.js in mishoo UglifyJS 3.13.2 via the name variable in ast.js. NOTE: the vendor considers this an invalid report.
<p>Publish Date: 2022-10-20
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-37598>CVE-2022-37598</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-10-20</p>
<p>Fix Resolution (uglify-js): 3.13.10</p>
<p>Direct dependency fix Resolution (jest): 22.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_infrastructure
|
cve high detected in uglify js tgz cve high severity vulnerability vulnerable library uglify js tgz javascript parser mangler compressor and beautifier toolkit library home page a href path to dependency file package json path to vulnerable library node modules handlebars node modules uglify js package json dependency hierarchy jest tgz root library jest cli tgz istanbul api tgz istanbul reports tgz handlebars tgz x uglify js tgz vulnerable library found in base branch stable vulnerability details disputed prototype pollution vulnerability in function defnode in ast js in mishoo uglifyjs via the name variable in ast js note the vendor considers this an invalid report publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution uglify js direct dependency fix resolution jest step up your open source security game with mend
| 0
|
17,759
| 5,512,274,452
|
IssuesEvent
|
2017-03-17 08:54:24
|
akvo/akvo-flow-mobile
|
https://api.github.com/repos/akvo/akvo-flow-mobile
|
closed
|
Create 3 modules for the layers of the app
|
Code Refactoring New and noteworthy Ready for release
|
This is the basic "skeleton" for clean architecture
1) Presentation layer (ui, activitiesโฆ)
/app -already exists
2) Domain layer (business logic, interactors)
/domain
3) Data logic (repositories: apis, database, shared preferences)
/data
App logic and components will be refactored and/or moved later on progressively.
https://8thlight.com/blog/uncle-bob/2012/08/13/the-clean-architecture.html
|
1.0
|
Create 3 modules for the layers of the app - This is the basic "skeleton" for clean architecture
1) Presentation layer (ui, activitiesโฆ)
/app -already exists
2) Domain layer (business logic, interactors)
/domain
3) Data logic (repositories: apis, database, shared preferences)
/data
App logic and components will be refactored and/or moved later on progressively.
https://8thlight.com/blog/uncle-bob/2012/08/13/the-clean-architecture.html
|
non_infrastructure
|
create modules for the layers of the app this is the basic skeleton for clean architecture presentation layer ui activitiesโฆ app already exists domain layer business logic interactors domain data logic repositories apis database shared preferences data app logic and components will be refactored and or moved later on progressively
| 0
|
790,272
| 27,821,139,777
|
IssuesEvent
|
2023-03-19 08:47:43
|
AY2223S2-CS2113-T12-4/tp
|
https://api.github.com/repos/AY2223S2-CS2113-T12-4/tp
|
closed
|
[Enhancement] Setup checkstyle to enforce CS2113 Coding Style
|
type.Enhancement priority.Low
|
**Describe your proposed enhancement.**
Import the checkstyle config for CS2113 coding style from the [addressbook-level3 repo](https://github.com/se-edu/addressbook-level3/tree/master/config/checkstyle). Setup your IntelliJ to enforce the coding style using [this guide](https://se-
**Why should this enhancement be added? Explain its benefits/significance.**
Coding style is one of the graded components for this project, and it's much more efficient to fix coding style violations downstream than have to flag them manually during PR review.
|
1.0
|
[Enhancement] Setup checkstyle to enforce CS2113 Coding Style - **Describe your proposed enhancement.**
Import the checkstyle config for CS2113 coding style from the [addressbook-level3 repo](https://github.com/se-edu/addressbook-level3/tree/master/config/checkstyle). Setup your IntelliJ to enforce the coding style using [this guide](https://se-
**Why should this enhancement be added? Explain its benefits/significance.**
Coding style is one of the graded components for this project, and it's much more efficient to fix coding style violations downstream than have to flag them manually during PR review.
|
non_infrastructure
|
setup checkstyle to enforce coding style describe your proposed enhancement import the checkstyle config for coding style from the setup your intellij to enforce the coding style using why should this enhancement be added explain its benefits significance coding style is one of the graded components for this project and it s much more efficient to fix coding style violations downstream than have to flag them manually during pr review
| 0
|
24,376
| 17,150,614,689
|
IssuesEvent
|
2021-07-13 20:03:38
|
ZcashFoundation/zebra
|
https://api.github.com/repos/ZcashFoundation/zebra
|
closed
|
macOS CI builders sometimes hang for half an hour or more
|
A-infrastructure C-bug I-integration-fail P-Low S-needs-triage
|
## Motivation
Zebra's macOS CI has failed twice in the last 4 hours.
The builders just seem to hang for a long time.
Sometimes they start again:
```
Fri, 09 Jul 2021 00:29:27 GMT Running `rustc --crate-name zebra_network --edition=2018 zebra-network/src/lib.rs --error-format=json --json=diagnostic-rendered-ansi --emit=dep-info,link -C embed-bitcode=no -C split-debuginfo=unpacked -C debuginfo=2 --test -C metadata=1ac22abde7ec2a58 -C extra-filename=-1ac22abde7ec2a58 --out-dir /Users/runner/work/zebra/zebra/target/debug/deps -L dependency=/Users/runner/work/zebra/zebra/target/debug/deps --extern bitflags=/Users/runner/work/zebra/zebra/target/debug/deps/libbitflags-0fcba523abfe46f1.rlib --extern byteorder=/Users/runner/work/zebra/zebra/target/debug/deps/libbyteorder-3f09ed744f9cea66.rlib --extern bytes=/Users/runner/work/zebra/zebra/target/debug/deps/libbytes-a867826a4c1ee23d.rlib --extern chrono=/Users/runner/work/zebra/zebra/target/debug/deps/libchrono-716c0b3f512de7d3.rlib --extern futures=/Users/runner/work/zebra/zebra/target/debug/deps/libfutures-ad19f46de33cddd5.rlib --extern hex=/Users/runner/work/zebra/zebra/target/debug/deps/libhex-f769a19b4a1634e0.rlib --extern indexmap=/Users/runner/work/zebra/zebra/target/debug/deps/libindexmap-952abc14b515d4d9.rlib --extern lazy_static=/Users/runner/work/zebra/zebra/target/debug/deps/liblazy_static-345966943c543af5.rlib --extern metrics=/Users/runner/work/zebra/zebra/target/debug/deps/libmetrics-11d555c1fba0e38b.rlib --extern pin_project=/Users/runner/work/zebra/zebra/target/debug/deps/libpin_project-7a86cd5f00c10e0e.rlib --extern proptest=/Users/runner/work/zebra/zebra/target/debug/deps/libproptest-bb113a575e3f9f0c.rlib --extern proptest_derive=/Users/runner/work/zebra/zebra/target/debug/deps/libproptest_derive-027d0af2b8a526e4.dylib --extern rand=/Users/runner/work/zebra/zebra/target/debug/deps/librand-9f908d57422416d7.rlib --extern regex=/Users/runner/work/zebra/zebra/target/debug/deps/libregex-80f2c798dd8e3513.rlib --extern serde=/Users/runner/work/zebra/zebra/target/debug/deps/libserde-c3ddac4fefdc041c.rlib --extern thiserror=/Users/runner/work/zebra/zebra/target/debug/deps/libthiserror-aedc752c25818fd6.rlib --extern tokio=/Users/runner/work/zebra/zebra/target/debug/deps/libtokio-164f26971a0c8b79.rlib --extern tokio_util=/Users/runner/work/zebra/zebra/target/debug/deps/libtokio_util-0db763c52e67b5d0.rlib --extern toml=/Users/runner/work/zebra/zebra/target/debug/deps/libtoml-999e6f9cc5b5091c.rlib --extern tower=/Users/runner/work/zebra/zebra/target/debug/deps/libtower-d1f706c120caf475.rlib --extern tracing=/Users/runner/work/zebra/zebra/target/debug/deps/libtracing-f8eea80e2edb8b13.rlib --extern tracing_error=/Users/runner/work/zebra/zebra/target/debug/deps/libtracing_error-4c4be8ccabc2c6c1.rlib --extern tracing_futures=/Users/runner/work/zebra/zebra/target/debug/deps/libtracing_futures-c346ccbc9c6ee47e.rlib --extern zebra_chain=/Users/runner/work/zebra/zebra/target/debug/deps/libzebra_chain-fed83e7905ee99e2.rlib --extern zebra_test=/Users/runner/work/zebra/zebra/target/debug/deps/libzebra_test-9122886a2a3552a3.rlib -L native=/Users/runner/work/zebra/zebra/target/debug/build/secp256k1-sys-41e5fcc6521abb5a/out`
Fri, 09 Jul 2021 00:59:11 GMT Running `rustc --crate-name librocksdb_sys --edition=2018 /Users/runner/.cargo/registry/src/github.com-1ecc6299db9ec823/librocksdb-sys-6.17.3/src/lib.rs --error-format=json --json=diagnostic-rendered-ansi,artifacts --crate-type lib --emit=dep-info,metadata,link -C embed-bitcode=no -C split-debuginfo=unpacked -C debuginfo=2 --cfg 'feature="bzip2"' --cfg 'feature="default"' --cfg 'feature="lz4"' --cfg 'feature="snappy"' --cfg 'feature="static"' --cfg 'feature="zlib"' --cfg 'feature="zstd"' -C metadata=7f8e25ac55fc6c82 -C extra-filename=-7f8e25ac55fc6c82 --out-dir /Users/runner/work/zebra/zebra/target/debug/deps -L dependency=/Users/runner/work/zebra/zebra/target/debug/deps --extern libc=/Users/runner/work/zebra/zebra/target/debug/deps/liblibc-84b9a47d29395960.rmeta --cap-lints allow -L native=/Users/runner/work/zebra/zebra/target/debug/build/librocksdb-sys-dcfd52201e2345eb/out -L native=/Users/runner/work/zebra/zebra/target/debug/build/librocksdb-sys-dcfd52201e2345eb/out -L native=/Users/runner/work/zebra/zebra/target/debug/build/librocksdb-sys-dcfd52201e2345eb/out -L native=/Users/runner/work/zebra/zebra/target/debug/build/librocksdb-sys-dcfd52201e2345eb/out -L native=/Users/runner/work/zebra/zebra/target/debug/build/librocksdb-sys-dcfd52201e2345eb/out -L native=/Users/runner/work/zebra/zebra/target/debug/build/librocksdb-sys-dcfd52201e2345eb/out -l static=rocksdb -l c++ -l static=snappy -l c++ -l static=lz4 -l static=zstd -l static=z -l static=bz2`
```
https://github.com/ZcashFoundation/zebra/pull/2472/checks?check_run_id=3024421827#step:10:450
And sometimes there is no log output after the hang:
https://github.com/ZcashFoundation/zebra/runs/3023384611#step:9:12
|
1.0
|
macOS CI builders sometimes hang for half an hour or more - ## Motivation
Zebra's macOS CI has failed twice in the last 4 hours.
The builders just seem to hang for a long time.
Sometimes they start again:
```
Fri, 09 Jul 2021 00:29:27 GMT Running `rustc --crate-name zebra_network --edition=2018 zebra-network/src/lib.rs --error-format=json --json=diagnostic-rendered-ansi --emit=dep-info,link -C embed-bitcode=no -C split-debuginfo=unpacked -C debuginfo=2 --test -C metadata=1ac22abde7ec2a58 -C extra-filename=-1ac22abde7ec2a58 --out-dir /Users/runner/work/zebra/zebra/target/debug/deps -L dependency=/Users/runner/work/zebra/zebra/target/debug/deps --extern bitflags=/Users/runner/work/zebra/zebra/target/debug/deps/libbitflags-0fcba523abfe46f1.rlib --extern byteorder=/Users/runner/work/zebra/zebra/target/debug/deps/libbyteorder-3f09ed744f9cea66.rlib --extern bytes=/Users/runner/work/zebra/zebra/target/debug/deps/libbytes-a867826a4c1ee23d.rlib --extern chrono=/Users/runner/work/zebra/zebra/target/debug/deps/libchrono-716c0b3f512de7d3.rlib --extern futures=/Users/runner/work/zebra/zebra/target/debug/deps/libfutures-ad19f46de33cddd5.rlib --extern hex=/Users/runner/work/zebra/zebra/target/debug/deps/libhex-f769a19b4a1634e0.rlib --extern indexmap=/Users/runner/work/zebra/zebra/target/debug/deps/libindexmap-952abc14b515d4d9.rlib --extern lazy_static=/Users/runner/work/zebra/zebra/target/debug/deps/liblazy_static-345966943c543af5.rlib --extern metrics=/Users/runner/work/zebra/zebra/target/debug/deps/libmetrics-11d555c1fba0e38b.rlib --extern pin_project=/Users/runner/work/zebra/zebra/target/debug/deps/libpin_project-7a86cd5f00c10e0e.rlib --extern proptest=/Users/runner/work/zebra/zebra/target/debug/deps/libproptest-bb113a575e3f9f0c.rlib --extern proptest_derive=/Users/runner/work/zebra/zebra/target/debug/deps/libproptest_derive-027d0af2b8a526e4.dylib --extern rand=/Users/runner/work/zebra/zebra/target/debug/deps/librand-9f908d57422416d7.rlib --extern regex=/Users/runner/work/zebra/zebra/target/debug/deps/libregex-80f2c798dd8e3513.rlib --extern serde=/Users/runner/work/zebra/zebra/target/debug/deps/libserde-c3ddac4fefdc041c.rlib --extern thiserror=/Users/runner/work/zebra/zebra/target/debug/deps/libthiserror-aedc752c25818fd6.rlib --extern tokio=/Users/runner/work/zebra/zebra/target/debug/deps/libtokio-164f26971a0c8b79.rlib --extern tokio_util=/Users/runner/work/zebra/zebra/target/debug/deps/libtokio_util-0db763c52e67b5d0.rlib --extern toml=/Users/runner/work/zebra/zebra/target/debug/deps/libtoml-999e6f9cc5b5091c.rlib --extern tower=/Users/runner/work/zebra/zebra/target/debug/deps/libtower-d1f706c120caf475.rlib --extern tracing=/Users/runner/work/zebra/zebra/target/debug/deps/libtracing-f8eea80e2edb8b13.rlib --extern tracing_error=/Users/runner/work/zebra/zebra/target/debug/deps/libtracing_error-4c4be8ccabc2c6c1.rlib --extern tracing_futures=/Users/runner/work/zebra/zebra/target/debug/deps/libtracing_futures-c346ccbc9c6ee47e.rlib --extern zebra_chain=/Users/runner/work/zebra/zebra/target/debug/deps/libzebra_chain-fed83e7905ee99e2.rlib --extern zebra_test=/Users/runner/work/zebra/zebra/target/debug/deps/libzebra_test-9122886a2a3552a3.rlib -L native=/Users/runner/work/zebra/zebra/target/debug/build/secp256k1-sys-41e5fcc6521abb5a/out`
Fri, 09 Jul 2021 00:59:11 GMT Running `rustc --crate-name librocksdb_sys --edition=2018 /Users/runner/.cargo/registry/src/github.com-1ecc6299db9ec823/librocksdb-sys-6.17.3/src/lib.rs --error-format=json --json=diagnostic-rendered-ansi,artifacts --crate-type lib --emit=dep-info,metadata,link -C embed-bitcode=no -C split-debuginfo=unpacked -C debuginfo=2 --cfg 'feature="bzip2"' --cfg 'feature="default"' --cfg 'feature="lz4"' --cfg 'feature="snappy"' --cfg 'feature="static"' --cfg 'feature="zlib"' --cfg 'feature="zstd"' -C metadata=7f8e25ac55fc6c82 -C extra-filename=-7f8e25ac55fc6c82 --out-dir /Users/runner/work/zebra/zebra/target/debug/deps -L dependency=/Users/runner/work/zebra/zebra/target/debug/deps --extern libc=/Users/runner/work/zebra/zebra/target/debug/deps/liblibc-84b9a47d29395960.rmeta --cap-lints allow -L native=/Users/runner/work/zebra/zebra/target/debug/build/librocksdb-sys-dcfd52201e2345eb/out -L native=/Users/runner/work/zebra/zebra/target/debug/build/librocksdb-sys-dcfd52201e2345eb/out -L native=/Users/runner/work/zebra/zebra/target/debug/build/librocksdb-sys-dcfd52201e2345eb/out -L native=/Users/runner/work/zebra/zebra/target/debug/build/librocksdb-sys-dcfd52201e2345eb/out -L native=/Users/runner/work/zebra/zebra/target/debug/build/librocksdb-sys-dcfd52201e2345eb/out -L native=/Users/runner/work/zebra/zebra/target/debug/build/librocksdb-sys-dcfd52201e2345eb/out -l static=rocksdb -l c++ -l static=snappy -l c++ -l static=lz4 -l static=zstd -l static=z -l static=bz2`
```
https://github.com/ZcashFoundation/zebra/pull/2472/checks?check_run_id=3024421827#step:10:450
And sometimes there is no log output after the hang:
https://github.com/ZcashFoundation/zebra/runs/3023384611#step:9:12
|
infrastructure
|
macos ci builders sometimes hang for half an hour or more motivation zebra s macos ci has failed twice in the last hours the builders just seem to hang for a long time sometimes they start again fri jul gmt running rustc crate name zebra network edition zebra network src lib rs error format json json diagnostic rendered ansi emit dep info link c embed bitcode no c split debuginfo unpacked c debuginfo test c metadata c extra filename out dir users runner work zebra zebra target debug deps l dependency users runner work zebra zebra target debug deps extern bitflags users runner work zebra zebra target debug deps libbitflags rlib extern byteorder users runner work zebra zebra target debug deps libbyteorder rlib extern bytes users runner work zebra zebra target debug deps libbytes rlib extern chrono users runner work zebra zebra target debug deps libchrono rlib extern futures users runner work zebra zebra target debug deps libfutures rlib extern hex users runner work zebra zebra target debug deps libhex rlib extern indexmap users runner work zebra zebra target debug deps libindexmap rlib extern lazy static users runner work zebra zebra target debug deps liblazy static rlib extern metrics users runner work zebra zebra target debug deps libmetrics rlib extern pin project users runner work zebra zebra target debug deps libpin project rlib extern proptest users runner work zebra zebra target debug deps libproptest rlib extern proptest derive users runner work zebra zebra target debug deps libproptest derive dylib extern rand users runner work zebra zebra target debug deps librand rlib extern regex users runner work zebra zebra target debug deps libregex rlib extern serde users runner work zebra zebra target debug deps libserde rlib extern thiserror users runner work zebra zebra target debug deps libthiserror rlib extern tokio users runner work zebra zebra target debug deps libtokio rlib extern tokio util users runner work zebra zebra target debug deps libtokio util rlib extern toml users runner work zebra zebra target debug deps libtoml rlib extern tower users runner work zebra zebra target debug deps libtower rlib extern tracing users runner work zebra zebra target debug deps libtracing rlib extern tracing error users runner work zebra zebra target debug deps libtracing error rlib extern tracing futures users runner work zebra zebra target debug deps libtracing futures rlib extern zebra chain users runner work zebra zebra target debug deps libzebra chain rlib extern zebra test users runner work zebra zebra target debug deps libzebra test rlib l native users runner work zebra zebra target debug build sys out fri jul gmt running rustc crate name librocksdb sys edition users runner cargo registry src github com librocksdb sys src lib rs error format json json diagnostic rendered ansi artifacts crate type lib emit dep info metadata link c embed bitcode no c split debuginfo unpacked c debuginfo cfg feature cfg feature default cfg feature cfg feature snappy cfg feature static cfg feature zlib cfg feature zstd c metadata c extra filename out dir users runner work zebra zebra target debug deps l dependency users runner work zebra zebra target debug deps extern libc users runner work zebra zebra target debug deps liblibc rmeta cap lints allow l native users runner work zebra zebra target debug build librocksdb sys out l native users runner work zebra zebra target debug build librocksdb sys out l native users runner work zebra zebra target debug build librocksdb sys out l native users runner work zebra zebra target debug build librocksdb sys out l native users runner work zebra zebra target debug build librocksdb sys out l native users runner work zebra zebra target debug build librocksdb sys out l static rocksdb l c l static snappy l c l static l static zstd l static z l static and sometimes there is no log output after the hang
| 1
|
6,140
| 6,188,990,704
|
IssuesEvent
|
2017-07-04 11:44:57
|
qutebrowser/qutebrowser
|
https://api.github.com/repos/qutebrowser/qutebrowser
|
opened
|
Travis upgrades for v1.0
|
component: infrastructure priority: 1 - middle
|
With #2742, we'll also need various environment upgrades on Travis:
- [ ] Drop `py34-cov` environment which uses Travis' Ubuntu Trusty
- [ ] Introduce coverage in one of the other enviroments (probably with QtWebEngine and adjusted coverage pragmas)
- [ ] Update Debian Jessie to Debian Stretch
- [ ] Update Ubuntu Xenial to Ubuntu Zesty (17.04) or Artful (17.10) - though it's probably similar enough to Debian Stretch anyways (all Qt 5.7.1), so maybe just drop it.
- [ ] Check if we can use container environments (`sudo: false`) for all non-Docker ones
- [ ] Switch to `lang: python`
|
1.0
|
Travis upgrades for v1.0 - With #2742, we'll also need various environment upgrades on Travis:
- [ ] Drop `py34-cov` environment which uses Travis' Ubuntu Trusty
- [ ] Introduce coverage in one of the other enviroments (probably with QtWebEngine and adjusted coverage pragmas)
- [ ] Update Debian Jessie to Debian Stretch
- [ ] Update Ubuntu Xenial to Ubuntu Zesty (17.04) or Artful (17.10) - though it's probably similar enough to Debian Stretch anyways (all Qt 5.7.1), so maybe just drop it.
- [ ] Check if we can use container environments (`sudo: false`) for all non-Docker ones
- [ ] Switch to `lang: python`
|
infrastructure
|
travis upgrades for with we ll also need various environment upgrades on travis drop cov environment which uses travis ubuntu trusty introduce coverage in one of the other enviroments probably with qtwebengine and adjusted coverage pragmas update debian jessie to debian stretch update ubuntu xenial to ubuntu zesty or artful though it s probably similar enough to debian stretch anyways all qt so maybe just drop it check if we can use container environments sudo false for all non docker ones switch to lang python
| 1
|
17,030
| 12,196,705,990
|
IssuesEvent
|
2020-04-29 19:28:10
|
SNLComputation/Albany
|
https://api.github.com/repos/SNLComputation/Albany
|
closed
|
consistent file extensions
|
infrastructure
|
Right now Albany uses a combination of `.h`, `.hpp`, `.cc`, and `.cpp` files. This makes writings scripts and doing general development more difficult. We should have all files be either `.hpp` or `.cpp`.
|
1.0
|
consistent file extensions - Right now Albany uses a combination of `.h`, `.hpp`, `.cc`, and `.cpp` files. This makes writings scripts and doing general development more difficult. We should have all files be either `.hpp` or `.cpp`.
|
infrastructure
|
consistent file extensions right now albany uses a combination of h hpp cc and cpp files this makes writings scripts and doing general development more difficult we should have all files be either hpp or cpp
| 1
|
4,020
| 4,771,709,856
|
IssuesEvent
|
2016-10-26 18:46:38
|
elmsln/elmsln
|
https://api.github.com/repos/elmsln/elmsln
|
closed
|
Support multiple versions of the same distro name space in same version
|
development infrastructure
|
#1406
We need to support multiple version of the same distro name space in the same stack. For example, cle-7.x-1.x and cle-7.x-2.x
would both get symlinked into stacks/studio/profiles/cle when it needs to be two different things
Proposal:
- profilename_versionunmber
- modify job file on creation for studio to add _version next to it
|
1.0
|
Support multiple versions of the same distro name space in same version - #1406
We need to support multiple version of the same distro name space in the same stack. For example, cle-7.x-1.x and cle-7.x-2.x
would both get symlinked into stacks/studio/profiles/cle when it needs to be two different things
Proposal:
- profilename_versionunmber
- modify job file on creation for studio to add _version next to it
|
infrastructure
|
support multiple versions of the same distro name space in same version we need to support multiple version of the same distro name space in the same stack for example cle x x and cle x x would both get symlinked into stacks studio profiles cle when it needs to be two different things proposal profilename versionunmber modify job file on creation for studio to add version next to it
| 1
|
21,597
| 14,660,469,416
|
IssuesEvent
|
2020-12-28 23:48:51
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
closed
|
Remove netstandard1.1 build configuration from System.Reflection.Context and harvest them instead when package 4.6.0 is released
|
area-Infrastructure-libraries enhancement
|
Milestone: Future --> after netcoreapp3.0 is released and the System.Reflection.Context package with version 4.6.0 is released.
|
1.0
|
Remove netstandard1.1 build configuration from System.Reflection.Context and harvest them instead when package 4.6.0 is released - Milestone: Future --> after netcoreapp3.0 is released and the System.Reflection.Context package with version 4.6.0 is released.
|
infrastructure
|
remove build configuration from system reflection context and harvest them instead when package is released milestone future after is released and the system reflection context package with version is released
| 1
|
10,722
| 15,766,180,886
|
IssuesEvent
|
2021-03-31 14:49:46
|
renovatebot/renovate
|
https://api.github.com/repos/renovatebot/renovate
|
closed
|
ignore versions gradle
|
priority-5-triage status:requirements type:feature
|
I'm trying to set up renovate (self-hosted) to ignore versions that have the suffix -dev, .dev, -rc, but I haven't been able to ignore it yet, it has brought versions dev, (1.1.0 to 1.4.1-dev )
I would like to know if how to configure renovate.json for it to ignore these versions
Could you write an example for me to follow?
|
1.0
|
ignore versions gradle - I'm trying to set up renovate (self-hosted) to ignore versions that have the suffix -dev, .dev, -rc, but I haven't been able to ignore it yet, it has brought versions dev, (1.1.0 to 1.4.1-dev )
I would like to know if how to configure renovate.json for it to ignore these versions
Could you write an example for me to follow?
|
non_infrastructure
|
ignore versions gradle i m trying to set up renovate self hosted to ignore versions that have the suffix dev dev rc but i haven t been able to ignore it yet it has brought versions dev to dev i would like to know if how to configure renovate json for it to ignore these versions could you write an example for me to follow
| 0
|
21,658
| 14,699,699,004
|
IssuesEvent
|
2021-01-04 09:00:56
|
filecoin-project/venus
|
https://api.github.com/repos/filecoin-project/venus
|
closed
|
create startup script to run through all steps needed before connecting to nightly cluster
|
A-infrastructure C-ux E-help-wanted P3
|
### Description
Create a simple startup script that folks new to Filecoin can run through to reconnect to the nightly cluster every day.
Should include the following:
```
cd $GOPATH/src/github.com/filecoin-project/go-filecoin
git pull origin master
go run ./build/*.go deps
go run ./build/*.go build
go run ./build/*.go install
rm -rf ~/.filecoin
go-filecoin init --cluster-nightly --genesisfile=http://nightly.kittyhawk.wtf:8020/genesis.car
go-filecoin daemon
```
Please create this script and add it to top-level directory of go-filecoin so it can be run easily from a user's CLI!
### Acceptance criteria
Reduce the typing of 8 (sometimes unwieldy) commands to one. Should be run like: `./startup.sh`
|
1.0
|
create startup script to run through all steps needed before connecting to nightly cluster - ### Description
Create a simple startup script that folks new to Filecoin can run through to reconnect to the nightly cluster every day.
Should include the following:
```
cd $GOPATH/src/github.com/filecoin-project/go-filecoin
git pull origin master
go run ./build/*.go deps
go run ./build/*.go build
go run ./build/*.go install
rm -rf ~/.filecoin
go-filecoin init --cluster-nightly --genesisfile=http://nightly.kittyhawk.wtf:8020/genesis.car
go-filecoin daemon
```
Please create this script and add it to top-level directory of go-filecoin so it can be run easily from a user's CLI!
### Acceptance criteria
Reduce the typing of 8 (sometimes unwieldy) commands to one. Should be run like: `./startup.sh`
|
infrastructure
|
create startup script to run through all steps needed before connecting to nightly cluster description create a simple startup script that folks new to filecoin can run through to reconnect to the nightly cluster every day should include the following cd gopath src github com filecoin project go filecoin git pull origin master go run build go deps go run build go build go run build go install rm rf filecoin go filecoin init cluster nightly genesisfile go filecoin daemon please create this script and add it to top level directory of go filecoin so it can be run easily from a user s cli acceptance criteria reduce the typing of sometimes unwieldy commands to one should be run like startup sh
| 1
|
5,393
| 5,633,813,020
|
IssuesEvent
|
2017-04-05 19:48:25
|
MIPT-ILab/mipt-mips
|
https://api.github.com/repos/MIPT-ILab/mipt-mips
|
closed
|
Add root Makefile to keep all compiler/linker options
|
1 point infrastructure
|
Currently we define CXXFLAGS in each Makefile. The object is to add one Makefile in $TRUNK to keep all the options.
**AMB:** root Makefile is added, all other Makefiles include it,
|
1.0
|
Add root Makefile to keep all compiler/linker options - Currently we define CXXFLAGS in each Makefile. The object is to add one Makefile in $TRUNK to keep all the options.
**AMB:** root Makefile is added, all other Makefiles include it,
|
infrastructure
|
add root makefile to keep all compiler linker options currently we define cxxflags in each makefile the object is to add one makefile in trunk to keep all the options amb root makefile is added all other makefiles include it
| 1
|
33,341
| 27,394,658,000
|
IssuesEvent
|
2023-02-28 18:42:54
|
ocaml/ocaml.org
|
https://api.github.com/repos/ocaml/ocaml.org
|
closed
|
Hosting a sourcehut instance
|
infrastructure suggestion
|
[Sourcehut](https://sourcehut.org) was requested by [some users](https://discuss.ocaml.org/t/ocaml-org-recapping-2022-and-queries-on-the-fediverse).
I don't know the process to get an instance running on `ocaml.org` so maybe someone who knows about it could explain what should be done first ?
|
1.0
|
Hosting a sourcehut instance - [Sourcehut](https://sourcehut.org) was requested by [some users](https://discuss.ocaml.org/t/ocaml-org-recapping-2022-and-queries-on-the-fediverse).
I don't know the process to get an instance running on `ocaml.org` so maybe someone who knows about it could explain what should be done first ?
|
infrastructure
|
hosting a sourcehut instance was requested by i don t know the process to get an instance running on ocaml org so maybe someone who knows about it could explain what should be done first
| 1
|
681,497
| 23,313,001,507
|
IssuesEvent
|
2022-08-08 09:56:24
|
steedos/steedos-platform
|
https://api.github.com/repos/steedos/steedos-platform
|
closed
|
ไผๅ๏ผlookupไธๆๆจกๅผโๆฐๅปบ**โๆ้ฎๅ่ฝ
|
done priority: High
|
lookupไธๆๆจกๅผ็นๅปโๆฐๅปบ**โๆ้ฎๅๅปบ่ฎฐๅฝ๏ผๅฆๆ้
็ฝฎไบreference_to_field๏ผ็นๅปโ็กฎ่ฎคโๅๆชๆพ็คบlabelใ
<img width="296" alt="image" src="https://user-images.githubusercontent.com/41402189/183056142-456a69bc-96d2-4f6f-a4fa-35f53e9bbc41.png">
|
1.0
|
ไผๅ๏ผlookupไธๆๆจกๅผโๆฐๅปบ**โๆ้ฎๅ่ฝ - lookupไธๆๆจกๅผ็นๅปโๆฐๅปบ**โๆ้ฎๅๅปบ่ฎฐๅฝ๏ผๅฆๆ้
็ฝฎไบreference_to_field๏ผ็นๅปโ็กฎ่ฎคโๅๆชๆพ็คบlabelใ
<img width="296" alt="image" src="https://user-images.githubusercontent.com/41402189/183056142-456a69bc-96d2-4f6f-a4fa-35f53e9bbc41.png">
|
non_infrastructure
|
ไผๅ๏ผlookupไธๆๆจกๅผโๆฐๅปบ โๆ้ฎๅ่ฝ lookupไธๆๆจกๅผ็นๅปโๆฐๅปบ โๆ้ฎๅๅปบ่ฎฐๅฝ๏ผๅฆๆ้
็ฝฎไบreference to field๏ผ็นๅปโ็กฎ่ฎคโๅๆชๆพ็คบlabelใ img width alt image src
| 0
|
177,808
| 21,509,187,245
|
IssuesEvent
|
2022-04-28 01:13:57
|
rgordon95/advanced-react-demo
|
https://api.github.com/repos/rgordon95/advanced-react-demo
|
closed
|
WS-2019-0318 (High) detected in handlebars-4.1.2.tgz - autoclosed
|
security vulnerability
|
## WS-2019-0318 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.1.2.tgz</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.1.2.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.1.2.tgz</a></p>
<p>Path to dependency file: /advanced-react-demo/package.json</p>
<p>Path to vulnerable library: advanced-react-demo/node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- jest-20.0.4.tgz (Root Library)
- jest-cli-20.0.4.tgz
- istanbul-api-1.3.7.tgz
- istanbul-reports-1.5.1.tgz
- :x: **handlebars-4.1.2.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In "showdownjs/showdown", versions prior to v4.4.5 are vulnerable against Regular expression Denial of Service (ReDOS) once receiving specially-crafted templates.
<p>Publish Date: 2019-10-20
<p>URL: <a href=https://github.com/wycats/handlebars.js/commit/8d5530ee2c3ea9f0aee3fde310b9f36887d00b8b>WS-2019-0318</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1300">https://www.npmjs.com/advisories/1300</a></p>
<p>Release Date: 2019-12-01</p>
<p>Fix Resolution: handlebars - 4.4.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
WS-2019-0318 (High) detected in handlebars-4.1.2.tgz - autoclosed - ## WS-2019-0318 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.1.2.tgz</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.1.2.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.1.2.tgz</a></p>
<p>Path to dependency file: /advanced-react-demo/package.json</p>
<p>Path to vulnerable library: advanced-react-demo/node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- jest-20.0.4.tgz (Root Library)
- jest-cli-20.0.4.tgz
- istanbul-api-1.3.7.tgz
- istanbul-reports-1.5.1.tgz
- :x: **handlebars-4.1.2.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In "showdownjs/showdown", versions prior to v4.4.5 are vulnerable against Regular expression Denial of Service (ReDOS) once receiving specially-crafted templates.
<p>Publish Date: 2019-10-20
<p>URL: <a href=https://github.com/wycats/handlebars.js/commit/8d5530ee2c3ea9f0aee3fde310b9f36887d00b8b>WS-2019-0318</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1300">https://www.npmjs.com/advisories/1300</a></p>
<p>Release Date: 2019-12-01</p>
<p>Fix Resolution: handlebars - 4.4.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_infrastructure
|
ws high detected in handlebars tgz autoclosed ws high severity vulnerability vulnerable library handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file advanced react demo package json path to vulnerable library advanced react demo node modules handlebars package json dependency hierarchy jest tgz root library jest cli tgz istanbul api tgz istanbul reports tgz x handlebars tgz vulnerable library vulnerability details in showdownjs showdown versions prior to are vulnerable against regular expression denial of service redos once receiving specially crafted templates publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution handlebars step up your open source security game with whitesource
| 0
|
10,575
| 4,074,267,061
|
IssuesEvent
|
2016-05-28 09:54:47
|
SleepyTrousers/EnderIO
|
https://api.github.com/repos/SleepyTrousers/EnderIO
|
closed
|
EnderIO-1.7.10-2.3.0.372_beta.jar crash
|
bug Code Complete
|
EnderIO-1.7.10-2.3.0.372_beta.jar
Description: Ticking block entity
blockVatTileEntity
http://pastebin.com/dtLdGJWm
EnderIO-1.7.10-2.3.0.371_beta.jar - stable and works fine
|
1.0
|
EnderIO-1.7.10-2.3.0.372_beta.jar crash - EnderIO-1.7.10-2.3.0.372_beta.jar
Description: Ticking block entity
blockVatTileEntity
http://pastebin.com/dtLdGJWm
EnderIO-1.7.10-2.3.0.371_beta.jar - stable and works fine
|
non_infrastructure
|
enderio beta jar crash enderio beta jar description ticking block entity blockvattileentity enderio beta jar stable and works fine
| 0
|
396,244
| 27,108,649,091
|
IssuesEvent
|
2023-02-15 13:56:08
|
cmendible/azqr
|
https://api.github.com/repos/cmendible/azqr
|
closed
|
Add guidance on how to install azqr on Azure Cloud Shell
|
documentation
|
## Describe the feature
Add guidance on how to install `azqr` on Azure Cloud Shell
|
1.0
|
Add guidance on how to install azqr on Azure Cloud Shell - ## Describe the feature
Add guidance on how to install `azqr` on Azure Cloud Shell
|
non_infrastructure
|
add guidance on how to install azqr on azure cloud shell describe the feature add guidance on how to install azqr on azure cloud shell
| 0
|
450,284
| 31,895,661,846
|
IssuesEvent
|
2023-09-18 01:15:39
|
ChewKeanHo/AutomataCI
|
https://api.github.com/repos/ChewKeanHo/AutomataCI
|
opened
|
Add Contributing Guidelines into Engineering Spec
|
Documentation Discussion
|
### Description
So that others know how to go about that `automataCI` directory.
### Page URL Address
_No response_
### Expected Presentation
There is a clear storyline flow of contributing guideline.
### Current Behavior
Contributing specs are missing.
### Associated Data Files
_No response_
|
1.0
|
Add Contributing Guidelines into Engineering Spec - ### Description
So that others know how to go about that `automataCI` directory.
### Page URL Address
_No response_
### Expected Presentation
There is a clear storyline flow of contributing guideline.
### Current Behavior
Contributing specs are missing.
### Associated Data Files
_No response_
|
non_infrastructure
|
add contributing guidelines into engineering spec description so that others know how to go about that automataci directory page url address no response expected presentation there is a clear storyline flow of contributing guideline current behavior contributing specs are missing associated data files no response
| 0
|
27,024
| 21,033,442,951
|
IssuesEvent
|
2022-03-31 04:39:08
|
zer0Kerbal/SimpleConstruction
|
https://api.github.com/repos/zer0Kerbal/SimpleConstruction
|
closed
|
[Bug ๐]: Metal Containers are not working for me
|
bug ๐ issue: config type: infrastructure type: fix work: complicated
|
### Brief description of your issue
None of the metal containers have any storage space. Metal doesn't even show up.
I'm not sure if this is a bug or because I was messing with the `ScienceLabs.cfg` file. (Removed some of the metal code then put it back)
### Steps to reproduce
If this is an issue just on my system, is there a way to completely reset the metal storage parts?
Removing the plugin doesn't seem to help.
### Expected behavior
Metal containers should have metal storage
### Actual behavior
No storage shows up for metal
### Environment
```shell
Mod: 4.0.99.4
KSP: 1.12.3.3173
Install: CurseForge App
[KSP.log](https://gist.github.com/DarinDev1000/f961ba66f2c8b1b8f31c21ebb1d2ebdf)
```
[KSP.log](https://gist.github.com/DarinDev1000/f961ba66f2c8b1b8f31c21ebb1d2ebdf)
### How did you download and install this?
OverWolf/CurseForge App (auto install)
|
1.0
|
[Bug ๐]: Metal Containers are not working for me - ### Brief description of your issue
None of the metal containers have any storage space. Metal doesn't even show up.
I'm not sure if this is a bug or because I was messing with the `ScienceLabs.cfg` file. (Removed some of the metal code then put it back)
### Steps to reproduce
If this is an issue just on my system, is there a way to completely reset the metal storage parts?
Removing the plugin doesn't seem to help.
### Expected behavior
Metal containers should have metal storage
### Actual behavior
No storage shows up for metal
### Environment
```shell
Mod: 4.0.99.4
KSP: 1.12.3.3173
Install: CurseForge App
[KSP.log](https://gist.github.com/DarinDev1000/f961ba66f2c8b1b8f31c21ebb1d2ebdf)
```
[KSP.log](https://gist.github.com/DarinDev1000/f961ba66f2c8b1b8f31c21ebb1d2ebdf)
### How did you download and install this?
OverWolf/CurseForge App (auto install)
|
infrastructure
|
metal containers are not working for me brief description of your issue none of the metal containers have any storage space metal doesn t even show up i m not sure if this is a bug or because i was messing with the sciencelabs cfg file removed some of the metal code then put it back steps to reproduce if this is an issue just on my system is there a way to completely reset the metal storage parts removing the plugin doesn t seem to help expected behavior metal containers should have metal storage actual behavior no storage shows up for metal environment shell mod ksp install curseforge app how did you download and install this overwolf curseforge app auto install
| 1
|
766,757
| 26,897,516,630
|
IssuesEvent
|
2023-02-06 13:30:54
|
googleapis/gaxios
|
https://api.github.com/repos/googleapis/gaxios
|
opened
|
Feature request: optional error throwing
|
type: feature request priority: p3
|
Hi, I recently came accustomed to using `Gaxios` through `googleapis` (specifically `drive_v3`). Seems all Promises are wrapped as GaxiosPromises which is okay but I have a minor nitpick to make.
While it's the norm to throw errors in JS I think it's mostly bad practise as this will often just crash the app and you can't type the errors as you could with normal returns.
I myself have been using this utility type:
```ts
export type Ok<T> = {
data: T
}
export type Error = {
err: string
code: number
}
export type Maybe<T> = Ok<T> | Error
```
to wrap all my error-producing code (mostly asynchronous) to gain full type-safety and proper error messages - akin to Go. I can then use them as such:
```ts
const resp = await driveService.listDrives(client);
if ("err" in resp) {
return next(new CustomError(resp.err, resp.code));
}
res.json(resp.data);
```
Now, I'm not suggesting you have to follow this schema (I used to use `ok: boolean` to check for errors but scrapped it once I saw `'data' in resp` works as well). But since you are wrapping Promises with your own type which includes both, success and error types, I'd love to be able to use those directly instead of catching the error, typing it manually and then doing my own thing on top of it causing just a lot boilerplate.
Ideally, there was a parameter / wrapper / utility method to allow returning an optional type which is either `GaxiosResponse<T> | GaxiosError`. Then I could myself check that for errors and not worry about typings and such.
|
1.0
|
Feature request: optional error throwing - Hi, I recently came accustomed to using `Gaxios` through `googleapis` (specifically `drive_v3`). Seems all Promises are wrapped as GaxiosPromises which is okay but I have a minor nitpick to make.
While it's the norm to throw errors in JS I think it's mostly bad practise as this will often just crash the app and you can't type the errors as you could with normal returns.
I myself have been using this utility type:
```ts
export type Ok<T> = {
data: T
}
export type Error = {
err: string
code: number
}
export type Maybe<T> = Ok<T> | Error
```
to wrap all my error-producing code (mostly asynchronous) to gain full type-safety and proper error messages - akin to Go. I can then use them as such:
```ts
const resp = await driveService.listDrives(client);
if ("err" in resp) {
return next(new CustomError(resp.err, resp.code));
}
res.json(resp.data);
```
Now, I'm not suggesting you have to follow this schema (I used to use `ok: boolean` to check for errors but scrapped it once I saw `'data' in resp` works as well). But since you are wrapping Promises with your own type which includes both, success and error types, I'd love to be able to use those directly instead of catching the error, typing it manually and then doing my own thing on top of it causing just a lot boilerplate.
Ideally, there was a parameter / wrapper / utility method to allow returning an optional type which is either `GaxiosResponse<T> | GaxiosError`. Then I could myself check that for errors and not worry about typings and such.
|
non_infrastructure
|
feature request optional error throwing hi i recently came accustomed to using gaxios through googleapis specifically drive seems all promises are wrapped as gaxiospromises which is okay but i have a minor nitpick to make while it s the norm to throw errors in js i think it s mostly bad practise as this will often just crash the app and you can t type the errors as you could with normal returns i myself have been using this utility type ts export type ok data t export type error err string code number export type maybe ok error to wrap all my error producing code mostly asynchronous to gain full type safety and proper error messages akin to go i can then use them as such ts const resp await driveservice listdrives client if err in resp return next new customerror resp err resp code res json resp data now i m not suggesting you have to follow this schema i used to use ok boolean to check for errors but scrapped it once i saw data in resp works as well but since you are wrapping promises with your own type which includes both success and error types i d love to be able to use those directly instead of catching the error typing it manually and then doing my own thing on top of it causing just a lot boilerplate ideally there was a parameter wrapper utility method to allow returning an optional type which is either gaxiosresponse gaxioserror then i could myself check that for errors and not worry about typings and such
| 0
|
143,938
| 22,205,532,780
|
IssuesEvent
|
2022-06-07 14:36:16
|
patternfly/patternfly-org
|
https://api.github.com/repos/patternfly/patternfly-org
|
opened
|
Create short guidelines for menu toggle component
|
PF4 design Guidelines
|
https://www.patternfly.org/v4/components/menu-toggle/design-guidelines
Could just be a couple sentences on what a menu toggle is, and when to use it (vs a dropdown even)
|
1.0
|
Create short guidelines for menu toggle component - https://www.patternfly.org/v4/components/menu-toggle/design-guidelines
Could just be a couple sentences on what a menu toggle is, and when to use it (vs a dropdown even)
|
non_infrastructure
|
create short guidelines for menu toggle component could just be a couple sentences on what a menu toggle is and when to use it vs a dropdown even
| 0
|
2,059
| 3,488,885,527
|
IssuesEvent
|
2016-01-03 11:34:31
|
asciidoctor/asciidoctor.js
|
https://api.github.com/repos/asciidoctor/asciidoctor.js
|
closed
|
Remove task runner (Grunt)
|
infrastructure
|
Task runner (Grunt, Gulp...) relies on plugins. This is good but also bad because only a few of them are actively maintained.
http://blog.keithcirkel.co.uk/why-we-should-stop-using-grunt/
I think we should just use npm as a task runner and remove Grunt:
http://blog.teamtreehouse.com/use-npm-task-runner
http://blog.keithcirkel.co.uk/how-to-use-npm-as-a-build-tool/
What do you think ?
|
1.0
|
Remove task runner (Grunt) - Task runner (Grunt, Gulp...) relies on plugins. This is good but also bad because only a few of them are actively maintained.
http://blog.keithcirkel.co.uk/why-we-should-stop-using-grunt/
I think we should just use npm as a task runner and remove Grunt:
http://blog.teamtreehouse.com/use-npm-task-runner
http://blog.keithcirkel.co.uk/how-to-use-npm-as-a-build-tool/
What do you think ?
|
infrastructure
|
remove task runner grunt task runner grunt gulp relies on plugins this is good but also bad because only a few of them are actively maintained i think we should just use npm as a task runner and remove grunt what do you think
| 1
|
30,517
| 24,883,555,217
|
IssuesEvent
|
2022-10-28 05:01:30
|
avniproject/avni-product
|
https://api.github.com/repos/avniproject/avni-product
|
closed
|
[Documentation] Create a page each for customers and users in the documentation site
|
Backlog Infrastructure/other Analysis Complete Story
|
For Customers
- Details to be provided for app users
- Details to be provided for dashboard users
- Details to be provided for theย users who will be logging the support tickets
- How to raise aย support ticket
- FAQ for dashboard
- FAQ for app
For Dashboard Users
- URL
- FAQ for dashboard
For App Users
- URL/Keyword to playstore
- FAQ
|
1.0
|
[Documentation] Create a page each for customers and users in the documentation site - For Customers
- Details to be provided for app users
- Details to be provided for dashboard users
- Details to be provided for theย users who will be logging the support tickets
- How to raise aย support ticket
- FAQ for dashboard
- FAQ for app
For Dashboard Users
- URL
- FAQ for dashboard
For App Users
- URL/Keyword to playstore
- FAQ
|
infrastructure
|
create a page each for customers and users in the documentation site for customers details to be provided for app users details to be provided for dashboard users details to be provided for theย users who will be logging the support tickets how to raise aย support ticket faq for dashboard faq for app for dashboard users url faq for dashboard for app users url keyword to playstore faq
| 1
|
30,372
| 24,765,423,668
|
IssuesEvent
|
2022-10-22 13:07:13
|
bounswe/bounswe2022group7
|
https://api.github.com/repos/bounswe/bounswe2022group7
|
closed
|
Backend Dockerization and Deployment to AWS
|
Status: Pending Review Priority: High Difficulty: Hard Target: Backend Target: Infrastructure
|
After [creating an initial project for backend](https://github.com/bounswe/bounswe2022group7/issues/251), we are to deploy it to AWS using dockerization.
**Reviewer:** @demet47
**Deadline**: 18/10/2022
|
1.0
|
Backend Dockerization and Deployment to AWS - After [creating an initial project for backend](https://github.com/bounswe/bounswe2022group7/issues/251), we are to deploy it to AWS using dockerization.
**Reviewer:** @demet47
**Deadline**: 18/10/2022
|
infrastructure
|
backend dockerization and deployment to aws after we are to deploy it to aws using dockerization reviewer deadline
| 1
|
1,294
| 3,122,742,359
|
IssuesEvent
|
2015-09-06 20:27:33
|
rexim/pacman.el
|
https://api.github.com/repos/rexim/pacman.el
|
closed
|
Emacs from EVM cannot recognize png images
|
bug infrastructure wontfix
|
https://travis-ci.org/rexim/pacman.el/jobs/77326855#L1666
The problem is explained in [this comment](https://github.com/rexim/pacman.el/issues/33#issuecomment-138117848)
Derived from #33
|
1.0
|
Emacs from EVM cannot recognize png images - https://travis-ci.org/rexim/pacman.el/jobs/77326855#L1666
The problem is explained in [this comment](https://github.com/rexim/pacman.el/issues/33#issuecomment-138117848)
Derived from #33
|
infrastructure
|
emacs from evm cannot recognize png images the problem is explained in derived from
| 1
|
22,456
| 15,207,888,216
|
IssuesEvent
|
2021-02-17 01:15:32
|
dotnet/roslyn
|
https://api.github.com/repos/dotnet/roslyn
|
closed
|
CI runs fail with "Logs already exists for build"
|
Area-Infrastructure Bug
|
Example run: https://dev.azure.com/dnceng/public/_build/results?buildId=877444&view=logs&j=7f45203e-c771-5e03-a313-4b8cfc15296a&t=9ba0a9f3-f509-5f5b-bb77-a8bffa790afb&l=28
I am wondering if it's impossible for test runs to succeed on retry sometimes because it can't upload logs at the end? I'm not sure if I'm interpreting the CI log properly or not.
/cc @jaredpar
|
1.0
|
CI runs fail with "Logs already exists for build" - Example run: https://dev.azure.com/dnceng/public/_build/results?buildId=877444&view=logs&j=7f45203e-c771-5e03-a313-4b8cfc15296a&t=9ba0a9f3-f509-5f5b-bb77-a8bffa790afb&l=28
I am wondering if it's impossible for test runs to succeed on retry sometimes because it can't upload logs at the end? I'm not sure if I'm interpreting the CI log properly or not.
/cc @jaredpar
|
infrastructure
|
ci runs fail with logs already exists for build example run i am wondering if it s impossible for test runs to succeed on retry sometimes because it can t upload logs at the end i m not sure if i m interpreting the ci log properly or not cc jaredpar
| 1
|
6,836
| 15,437,212,987
|
IssuesEvent
|
2021-03-07 15:54:07
|
tldr-pages/tldr
|
https://api.github.com/repos/tldr-pages/tldr
|
closed
|
Removing pages.pt_PT
|
architecture decision mass changes page edit
|
I would like to propose to merge `pages.pt_PT` into `pages.pt_BR` and the remove `pages.pt_PT`, and then rename `pages.pt_BR` โ `pages.pt` (maybe not rename it though, I'm not sure what effect that might have on clients ๐ค).
My reasoning is as follows:
- Notice the difference of `common/xkill.md`, for example:
```diff
> Termina o cliente associado a um elemento grรกfico.
> Utilizado para forรงar a terminaรงรฃo de processos que nรฃo respondem ou nรฃo apresentam botรฃo "fechar".
-- Ativar um cursor para fechar uma janela com o clique do botรฃo esquerdo do rato (pressionar qualquer outro botรฃo para
cancelar):
+- Ativar um cursor para fechar uma janela com o clique do botรฃo esquerdo do mouse (pressionar qualquer outro botรฃo
para cancelar):
`xkill`
```
- In `touch.md`, for example, every line is different between the two translations, but the only different is that `pt_PT` uses the noun `ficheiro` while `pt_BR` uses `arquivo`. (`file` vs. `archive`?)
- Besides those seemingly minor differences (which may just be wording usages on the part of the translator, not even necessarily vocabulary differences between the two dialects), there are **151** `pt_BR` pages and **9** `pt_PT` pages.
- There are other such dialectic differences with other languages, e.g. American English vs. British English, but we don't have a `pages.en_US` and `pages.en_GB`.
|
1.0
|
Removing pages.pt_PT - I would like to propose to merge `pages.pt_PT` into `pages.pt_BR` and the remove `pages.pt_PT`, and then rename `pages.pt_BR` โ `pages.pt` (maybe not rename it though, I'm not sure what effect that might have on clients ๐ค).
My reasoning is as follows:
- Notice the difference of `common/xkill.md`, for example:
```diff
> Termina o cliente associado a um elemento grรกfico.
> Utilizado para forรงar a terminaรงรฃo de processos que nรฃo respondem ou nรฃo apresentam botรฃo "fechar".
-- Ativar um cursor para fechar uma janela com o clique do botรฃo esquerdo do rato (pressionar qualquer outro botรฃo para
cancelar):
+- Ativar um cursor para fechar uma janela com o clique do botรฃo esquerdo do mouse (pressionar qualquer outro botรฃo
para cancelar):
`xkill`
```
- In `touch.md`, for example, every line is different between the two translations, but the only different is that `pt_PT` uses the noun `ficheiro` while `pt_BR` uses `arquivo`. (`file` vs. `archive`?)
- Besides those seemingly minor differences (which may just be wording usages on the part of the translator, not even necessarily vocabulary differences between the two dialects), there are **151** `pt_BR` pages and **9** `pt_PT` pages.
- There are other such dialectic differences with other languages, e.g. American English vs. British English, but we don't have a `pages.en_US` and `pages.en_GB`.
|
non_infrastructure
|
removing pages pt pt i would like to propose to merge pages pt pt into pages pt br and the remove pages pt pt and then rename pages pt br โ pages pt maybe not rename it though i m not sure what effect that might have on clients ๐ค my reasoning is as follows notice the difference of common xkill md for example diff termina o cliente associado a um elemento grรกfico utilizado para forรงar a terminaรงรฃo de processos que nรฃo respondem ou nรฃo apresentam botรฃo fechar ativar um cursor para fechar uma janela com o clique do botรฃo esquerdo do rato pressionar qualquer outro botรฃo para cancelar ativar um cursor para fechar uma janela com o clique do botรฃo esquerdo do mouse pressionar qualquer outro botรฃo para cancelar xkill in touch md for example every line is different between the two translations but the only different is that pt pt uses the noun ficheiro while pt br uses arquivo file vs archive besides those seemingly minor differences which may just be wording usages on the part of the translator not even necessarily vocabulary differences between the two dialects there are pt br pages and pt pt pages there are other such dialectic differences with other languages e g american english vs british english but we don t have a pages en us and pages en gb
| 0
|
236,607
| 26,031,788,916
|
IssuesEvent
|
2022-12-21 22:11:06
|
CMSgov/mdct-mcr
|
https://api.github.com/repos/CMSgov/mdct-mcr
|
closed
|
SHF - mdct-mcr - main - MEDIUM - S3.12 S3 access control lists (ACLs) should not be used to manage user access to buckets
|
security-hub main
|
**************************************************************
__This issue was generated from Security Hub data and is managed through automation.__
Please do not edit the title or body of this issue, or remove the security-hub tag. All other edits/comments are welcome.
Finding Id: arn:aws:securityhub:us-east-1:446712541566:subscription/aws-foundational-security-best-practices/v/1.0.0/S3.12/finding/42d369b3-ba1e-4d44-aadf-f5bcdc0a3c3a
**************************************************************
## Type of Issue:
- [x] Security Hub Finding
## Title:
S3.12 S3 access control lists (ACLs) should not be used to manage user access to buckets
## Id:
arn:aws:securityhub:us-east-1:446712541566:subscription/aws-foundational-security-best-practices/v/1.0.0/S3.12/finding/42d369b3-ba1e-4d44-aadf-f5bcdc0a3c3a
(You may use this ID to lookup this finding's details in Security Hub)
## Description
This control checks if S3 buckets allow user permissions via access control lists (ACLs). This control fails if ACLs are configured for user access on S3 Bucket.
## Remediation
https://docs.aws.amazon.com/console/securityhub/S3.12/remediation
## AC:
- The security hub finding is resolved or suppressed, indicated by a Workflow Status of Resolved or Suppressed.
|
True
|
SHF - mdct-mcr - main - MEDIUM - S3.12 S3 access control lists (ACLs) should not be used to manage user access to buckets - **************************************************************
__This issue was generated from Security Hub data and is managed through automation.__
Please do not edit the title or body of this issue, or remove the security-hub tag. All other edits/comments are welcome.
Finding Id: arn:aws:securityhub:us-east-1:446712541566:subscription/aws-foundational-security-best-practices/v/1.0.0/S3.12/finding/42d369b3-ba1e-4d44-aadf-f5bcdc0a3c3a
**************************************************************
## Type of Issue:
- [x] Security Hub Finding
## Title:
S3.12 S3 access control lists (ACLs) should not be used to manage user access to buckets
## Id:
arn:aws:securityhub:us-east-1:446712541566:subscription/aws-foundational-security-best-practices/v/1.0.0/S3.12/finding/42d369b3-ba1e-4d44-aadf-f5bcdc0a3c3a
(You may use this ID to lookup this finding's details in Security Hub)
## Description
This control checks if S3 buckets allow user permissions via access control lists (ACLs). This control fails if ACLs are configured for user access on S3 Bucket.
## Remediation
https://docs.aws.amazon.com/console/securityhub/S3.12/remediation
## AC:
- The security hub finding is resolved or suppressed, indicated by a Workflow Status of Resolved or Suppressed.
|
non_infrastructure
|
shf mdct mcr main medium access control lists acls should not be used to manage user access to buckets this issue was generated from security hub data and is managed through automation please do not edit the title or body of this issue or remove the security hub tag all other edits comments are welcome finding id arn aws securityhub us east subscription aws foundational security best practices v finding aadf type of issue security hub finding title access control lists acls should not be used to manage user access to buckets id arn aws securityhub us east subscription aws foundational security best practices v finding aadf you may use this id to lookup this finding s details in security hub description this control checks if buckets allow user permissions via access control lists acls this control fails if acls are configured for user access on bucket remediation ac the security hub finding is resolved or suppressed indicated by a workflow status of resolved or suppressed
| 0
|
749,120
| 26,150,216,714
|
IssuesEvent
|
2022-12-30 12:25:40
|
NomicFoundation/hardhat
|
https://api.github.com/repos/NomicFoundation/hardhat
|
closed
|
Support for yarn berry (v2)
|
priority:low
|
[Berry](https://github.com/yarnpkg/berry) is yarn's v2 release and is disruptive in many regards. They replaced "node_modules" with a new [plug-n-play](https://classic.yarnpkg.com/en/docs/pnp/) system that caches dependencies more efficiently.
In the docs for [Typescript support](https://buidler.dev/guides/typescript.html#plugin-type-extensions), you specify that we should add this to our `tsconfig.json` config:
```json
"files": [
"./buidler.config.ts",
"./node_modules/@nomiclabs/buidler-ethers/src/type-extensions.d.ts",
"./node_modules/@nomiclabs/buidler-waffle/src/type-extensions.d.ts"
]
```
But this is sadly not possible in yarn v2. I suppose that this is not a priority, but I wanted to open an issue about it anyways, in case more people stumble upon it and want to signal their support.
|
1.0
|
Support for yarn berry (v2) - [Berry](https://github.com/yarnpkg/berry) is yarn's v2 release and is disruptive in many regards. They replaced "node_modules" with a new [plug-n-play](https://classic.yarnpkg.com/en/docs/pnp/) system that caches dependencies more efficiently.
In the docs for [Typescript support](https://buidler.dev/guides/typescript.html#plugin-type-extensions), you specify that we should add this to our `tsconfig.json` config:
```json
"files": [
"./buidler.config.ts",
"./node_modules/@nomiclabs/buidler-ethers/src/type-extensions.d.ts",
"./node_modules/@nomiclabs/buidler-waffle/src/type-extensions.d.ts"
]
```
But this is sadly not possible in yarn v2. I suppose that this is not a priority, but I wanted to open an issue about it anyways, in case more people stumble upon it and want to signal their support.
|
non_infrastructure
|
support for yarn berry is yarn s release and is disruptive in many regards they replaced node modules with a new system that caches dependencies more efficiently in the docs for you specify that we should add this to our tsconfig json config json files buidler config ts node modules nomiclabs buidler ethers src type extensions d ts node modules nomiclabs buidler waffle src type extensions d ts but this is sadly not possible in yarn i suppose that this is not a priority but i wanted to open an issue about it anyways in case more people stumble upon it and want to signal their support
| 0
|
284,529
| 24,606,214,755
|
IssuesEvent
|
2022-10-14 16:28:01
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
opened
|
sql: truncate split unit test does not use a stable output and retries excessively
|
C-bug A-testing
|
**Describe the problem**
This logic test (in `crdb_internal`):
```
# Ensure that there are no longer any splits left over on the original indexes.
# TRUNCATE will have created equivalent splits points on the new indexes, so
# this test just checks that there are no more on the old indexes.
query TT retry
SELECT start_pretty, end_pretty FROM crdb_internal.ranges
WHERE split_enforced_until IS NOT NULL
AND (start_pretty LIKE '/Table/112/1%' OR start_pretty LIKE '/Table/112/2%')
----
/Table/112/1/1 /Table/112/1/2
/Table/112/1/2 /Table/112/1/3
/Table/112/1/3 /Table/112/2/1
/Table/112/2/1 /Table/112/2/2
/Table/112/2/2 /Table/112/2/3
/Table/112/2/3 /Table/112/3/1
```
Notice there is no `rowsort` nor `ORDER BY`. The output is unstable. Every time we use `--rewrite` we get a different reference output.
The reason why this does not appear to fail is because of the `retry` clause. But that behavior is severely problematic: it just makes the test re-run the query forever until the apparent order of the result rows matches the expected output.
Let's not do that and add the missing `rowsort` clause instead.
cc @jordanlewis @jeffrey-xiao @chengxiong-ruan for triage.
|
1.0
|
sql: truncate split unit test does not use a stable output and retries excessively - **Describe the problem**
This logic test (in `crdb_internal`):
```
# Ensure that there are no longer any splits left over on the original indexes.
# TRUNCATE will have created equivalent splits points on the new indexes, so
# this test just checks that there are no more on the old indexes.
query TT retry
SELECT start_pretty, end_pretty FROM crdb_internal.ranges
WHERE split_enforced_until IS NOT NULL
AND (start_pretty LIKE '/Table/112/1%' OR start_pretty LIKE '/Table/112/2%')
----
/Table/112/1/1 /Table/112/1/2
/Table/112/1/2 /Table/112/1/3
/Table/112/1/3 /Table/112/2/1
/Table/112/2/1 /Table/112/2/2
/Table/112/2/2 /Table/112/2/3
/Table/112/2/3 /Table/112/3/1
```
Notice there is no `rowsort` nor `ORDER BY`. The output is unstable. Every time we use `--rewrite` we get a different reference output.
The reason why this does not appear to fail is because of the `retry` clause. But that behavior is severely problematic: it just makes the test re-run the query forever until the apparent order of the result rows matches the expected output.
Let's not do that and add the missing `rowsort` clause instead.
cc @jordanlewis @jeffrey-xiao @chengxiong-ruan for triage.
|
non_infrastructure
|
sql truncate split unit test does not use a stable output and retries excessively describe the problem this logic test in crdb internal ensure that there are no longer any splits left over on the original indexes truncate will have created equivalent splits points on the new indexes so this test just checks that there are no more on the old indexes query tt retry select start pretty end pretty from crdb internal ranges where split enforced until is not null and start pretty like table or start pretty like table table table table table table table table table table table table table notice there is no rowsort nor order by the output is unstable every time we use rewrite we get a different reference output the reason why this does not appear to fail is because of the retry clause but that behavior is severely problematic it just makes the test re run the query forever until the apparent order of the result rows matches the expected output let s not do that and add the missing rowsort clause instead cc jordanlewis jeffrey xiao chengxiong ruan for triage
| 0
|
32,202
| 26,526,778,053
|
IssuesEvent
|
2023-01-19 09:22:27
|
TouchTheGrass/tpc-core
|
https://api.github.com/repos/TouchTheGrass/tpc-core
|
closed
|
ะะพะดะณะพัะพะฒะธัั ะธะฝััะฐััััะบัััั ะดะปั ะธัะฟะพะปัะทะพะฒะฐะฝะธั ะฒะตะฑ-ัะพะบะตัะพะฒ
|
infrastructure
|
ะขัะตะฑัะตััั ะฟะพะดะบะปััะธัั ะฑะธะฑะปะธะพัะตะบั `channels` ะดะปั ะธัะฟะพะปัะทะพะฒะฐะฝะธั ะฒะตะฑัะพะบะตัะพะฒ ัะตัะฒะตัะพะผ
|
1.0
|
ะะพะดะณะพัะพะฒะธัั ะธะฝััะฐััััะบัััั ะดะปั ะธัะฟะพะปัะทะพะฒะฐะฝะธั ะฒะตะฑ-ัะพะบะตัะพะฒ - ะขัะตะฑัะตััั ะฟะพะดะบะปััะธัั ะฑะธะฑะปะธะพัะตะบั `channels` ะดะปั ะธัะฟะพะปัะทะพะฒะฐะฝะธั ะฒะตะฑัะพะบะตัะพะฒ ัะตัะฒะตัะพะผ
|
infrastructure
|
ะฟะพะดะณะพัะพะฒะธัั ะธะฝััะฐััััะบัััั ะดะปั ะธัะฟะพะปัะทะพะฒะฐะฝะธั ะฒะตะฑ ัะพะบะตัะพะฒ ััะตะฑัะตััั ะฟะพะดะบะปััะธัั ะฑะธะฑะปะธะพัะตะบั channels ะดะปั ะธัะฟะพะปัะทะพะฒะฐะฝะธั ะฒะตะฑัะพะบะตัะพะฒ ัะตัะฒะตัะพะผ
| 1
|
4,616
| 5,199,016,132
|
IssuesEvent
|
2017-01-23 19:43:18
|
facebookincubator/create-react-app
|
https://api.github.com/repos/facebookincubator/create-react-app
|
closed
|
Use a more sophisticated template for end-to-end testing of Create React App
|
infrastructure
|
We currently test the same [`template`](https://github.com/facebookincubator/create-react-app/tree/master/packages/react-scripts/template) we generate by default, which is probably file, but we miss out on testing more advanced features. Ideally we should have a `kitchensink` template that uses all possible features (JSX, Flow, `NODE_PATH`, generators, async/await, class properties etc), and make sure that _that_ template builds and passes tests.
If youโd like to work on it, please write in this issue so others donโt duplicate effort. The task would consist of:
1. Looking at `tasks/e2e.sh` and learning how we test CRA
2. Figuring out how to make it use a different template just for e2e testโnot sure how Iโd do that so suggestions welcome
3. Creating that kitchensink template
|
1.0
|
Use a more sophisticated template for end-to-end testing of Create React App - We currently test the same [`template`](https://github.com/facebookincubator/create-react-app/tree/master/packages/react-scripts/template) we generate by default, which is probably file, but we miss out on testing more advanced features. Ideally we should have a `kitchensink` template that uses all possible features (JSX, Flow, `NODE_PATH`, generators, async/await, class properties etc), and make sure that _that_ template builds and passes tests.
If youโd like to work on it, please write in this issue so others donโt duplicate effort. The task would consist of:
1. Looking at `tasks/e2e.sh` and learning how we test CRA
2. Figuring out how to make it use a different template just for e2e testโnot sure how Iโd do that so suggestions welcome
3. Creating that kitchensink template
|
infrastructure
|
use a more sophisticated template for end to end testing of create react app we currently test the same we generate by default which is probably file but we miss out on testing more advanced features ideally we should have a kitchensink template that uses all possible features jsx flow node path generators async await class properties etc and make sure that that template builds and passes tests if youโd like to work on it please write in this issue so others donโt duplicate effort the task would consist of looking at tasks sh and learning how we test cra figuring out how to make it use a different template just for testโnot sure how iโd do that so suggestions welcome creating that kitchensink template
| 1
|
28,196
| 23,077,339,690
|
IssuesEvent
|
2022-07-26 01:49:57
|
m-labs/artiq
|
https://api.github.com/repos/m-labs/artiq
|
closed
|
artiq_flash: bail if scan chain is wrong
|
prio:2-normal area:coredevice area:infrastructure complexity:medium good first issue
|
Instead of continuing blindly, `artiq_flash` should have `openocd` error out if the JTAG scan chain is not as expected.
```
Design: top;COMPRESS=TRUE;UserID=0XFFFFFFFF;Version=2017.4
Part name: xcku040-ffva1156-1-c
Date: 2018/02/23
Time: 22:23:16
Bitstream payload length: 0x6d4708
Open On-Chip Debugger 0.10.0-00010-g9f8c46a (2018-01-30-14:40)
Licensed under GNU GPL v2
For bug reports, read
http://openocd.org/doc/doxygen/bugs.html
none separate
adapter speed: 5000 kHz
Info : clock speed 5000 kHz
Info : JTAG tap: xc7.tap tap/device found: 0x13631093 (mfg: 0x049 (Xilinx), part: 0x3631, ver: 0x1)
Info : JTAG tap: xcu.tap tap/device found: 0xffffffff (mfg: 0x7ff (<invalid>), part: 0xffff, ver: 0xf)
Warn : JTAG tap: xcu.tap UNEXPECTED: 0xffffffff (mfg: 0x7ff (<invalid>), part: 0xffff, ver: 0xf)
Error: JTAG tap: xcu.tap expected 1 of 1: 0x03822093 (mfg: 0x049 (Xilinx), part: 0x3822, ver: 0x0)
Error: Trying to use configured scan chain anyway...
Error: xcu.tap: IR capture error; saw 0x3f not 0x01
Warn : Bypassing JTAG setup events due to errors
Info : gdb server disabled
RTM FPGA XADC:
TEMP -273.15 C
VCCINT 0.000 V
VCCAUX 0.000 V
VCCBRAM 0.000 V
...
```
|
1.0
|
artiq_flash: bail if scan chain is wrong - Instead of continuing blindly, `artiq_flash` should have `openocd` error out if the JTAG scan chain is not as expected.
```
Design: top;COMPRESS=TRUE;UserID=0XFFFFFFFF;Version=2017.4
Part name: xcku040-ffva1156-1-c
Date: 2018/02/23
Time: 22:23:16
Bitstream payload length: 0x6d4708
Open On-Chip Debugger 0.10.0-00010-g9f8c46a (2018-01-30-14:40)
Licensed under GNU GPL v2
For bug reports, read
http://openocd.org/doc/doxygen/bugs.html
none separate
adapter speed: 5000 kHz
Info : clock speed 5000 kHz
Info : JTAG tap: xc7.tap tap/device found: 0x13631093 (mfg: 0x049 (Xilinx), part: 0x3631, ver: 0x1)
Info : JTAG tap: xcu.tap tap/device found: 0xffffffff (mfg: 0x7ff (<invalid>), part: 0xffff, ver: 0xf)
Warn : JTAG tap: xcu.tap UNEXPECTED: 0xffffffff (mfg: 0x7ff (<invalid>), part: 0xffff, ver: 0xf)
Error: JTAG tap: xcu.tap expected 1 of 1: 0x03822093 (mfg: 0x049 (Xilinx), part: 0x3822, ver: 0x0)
Error: Trying to use configured scan chain anyway...
Error: xcu.tap: IR capture error; saw 0x3f not 0x01
Warn : Bypassing JTAG setup events due to errors
Info : gdb server disabled
RTM FPGA XADC:
TEMP -273.15 C
VCCINT 0.000 V
VCCAUX 0.000 V
VCCBRAM 0.000 V
...
```
|
infrastructure
|
artiq flash bail if scan chain is wrong instead of continuing blindly artiq flash should have openocd error out if the jtag scan chain is not as expected design top compress true userid version part name c date time bitstream payload length open on chip debugger licensed under gnu gpl for bug reports read none separate adapter speed khz info clock speed khz info jtag tap tap tap device found mfg xilinx part ver info jtag tap xcu tap tap device found mfg part ver warn jtag tap xcu tap unexpected mfg part ver error jtag tap xcu tap expected of mfg xilinx part ver error trying to use configured scan chain anyway error xcu tap ir capture error saw not warn bypassing jtag setup events due to errors info gdb server disabled rtm fpga xadc temp c vccint v vccaux v vccbram v
| 1
|
611,848
| 18,982,824,589
|
IssuesEvent
|
2021-11-21 07:21:09
|
kimlimjustin/xplorer
|
https://api.github.com/repos/kimlimjustin/xplorer
|
closed
|
Current Implementation of Theme Plugin can Cause Critical Security Concern
|
bug priority: IMPORTANT
|
### Description
The theme customization allows user to configure the theme dynamically, however it also allow hackers to inject malicious code which may cause critical security issue.
### Steps To Reproduce
1. Edit the theme file (*.js)
2. Insert script such as `document.body.innerHTML = '';`
3. Run the script
### Expected behavior
Users are able to customize the theme (via another approaches) without having any security concerns.
### Xplorer Version
v0.1.0
### Operating System Version
Windows 11 Version 21H2 (Build 22000.258)
### Additional Information
It will be better if users could customize the theme via GUI like how windows 11 implement it.
|
1.0
|
Current Implementation of Theme Plugin can Cause Critical Security Concern - ### Description
The theme customization allows user to configure the theme dynamically, however it also allow hackers to inject malicious code which may cause critical security issue.
### Steps To Reproduce
1. Edit the theme file (*.js)
2. Insert script such as `document.body.innerHTML = '';`
3. Run the script
### Expected behavior
Users are able to customize the theme (via another approaches) without having any security concerns.
### Xplorer Version
v0.1.0
### Operating System Version
Windows 11 Version 21H2 (Build 22000.258)
### Additional Information
It will be better if users could customize the theme via GUI like how windows 11 implement it.
|
non_infrastructure
|
current implementation of theme plugin can cause critical security concern description the theme customization allows user to configure the theme dynamically however it also allow hackers to inject malicious code which may cause critical security issue steps to reproduce edit the theme file js insert script such as document body innerhtml run the script expected behavior users are able to customize the theme via another approaches without having any security concerns xplorer version operating system version windows version build additional information it will be better if users could customize the theme via gui like how windows implement it
| 0
|
31,205
| 25,412,335,315
|
IssuesEvent
|
2022-11-22 20:10:18
|
scikit-image/scikit-image
|
https://api.github.com/repos/scikit-image/scikit-image
|
closed
|
CI build failures on MacOS and Windows
|
:bug: Bug :robot: Infrastructure
|
### Description:
In #6621 and #6624, I see build failures on both Windows and MacOS. On Azure, this occurs when calling `python setup.py bdist_wheel` and seems [related to recent changes in pip 65.6](https://github.com/pypa/setuptools/issues/3693).
We had previously pinned to 65.5 in pyproject.toml in #6519, but that doesn't get used by the two cases above.
### Way to reproduce:
_No response_
### Traceback or output:
_No response_
### Version information:
_No response_
|
1.0
|
CI build failures on MacOS and Windows - ### Description:
In #6621 and #6624, I see build failures on both Windows and MacOS. On Azure, this occurs when calling `python setup.py bdist_wheel` and seems [related to recent changes in pip 65.6](https://github.com/pypa/setuptools/issues/3693).
We had previously pinned to 65.5 in pyproject.toml in #6519, but that doesn't get used by the two cases above.
### Way to reproduce:
_No response_
### Traceback or output:
_No response_
### Version information:
_No response_
|
infrastructure
|
ci build failures on macos and windows description in and i see build failures on both windows and macos on azure this occurs when calling python setup py bdist wheel and seems we had previously pinned to in pyproject toml in but that doesn t get used by the two cases above way to reproduce no response traceback or output no response version information no response
| 1
|
4,096
| 4,794,083,846
|
IssuesEvent
|
2016-10-31 20:00:13
|
eslint/eslint
|
https://api.github.com/repos/eslint/eslint
|
opened
|
Proposal: semver-minor and semver-patch labels
|
evaluating infrastructure
|
At the moment, we use the `enhancement` and `feature` labels for any semver-minor change. However, according to our [semver policy](https://github.com/eslint/eslint/blob/master/README.md#semantic-versioning-policy), not all semver-minor changes are enhancements; for example, fixing a false negative in a rule is a semver-minor change. We currently put the `enhancement` label on false-negative bugfixes to clarify that they're semver-minor, but I think this is mildly confusing; a false-negative bugfix isn't any more of an "enhancement" than a false-positive bugfix is. IMO, we're overloading the "enhancement" label to mean "semver-minor".
As a solution, I think we should add `semver-minor` and `semver-patch` labels. This would make it easier to see what can be merged for a patch release, and we would be able to use the `enhancement` label only for actual enhancements, such as new rule options.
---
I think our use `Fix:` and `Update:` in commit messages has the same problem; false-negative fixes use the `Update:` label to indicate that they're semver-minor. For example, some commit messages look like this:
```
Update: fix false negative of foo
```
But I think messages like this are easier to understand:
```
Fix: false negative of foo
```
The second commit message makes it easier for anyone skimming the changelog for new features (since they can just look at all commits with `Update:` and `New:` prefixes). At the moment, we use the `Update:` prefixes for bugfixes as well, which is a bit misleading to the reader.
(However, I know that these prefixes are used by the release tool to determine what kind of release to create, so I don't have a good solution for this problem at the moment aside from using separate prefixes for different types of bugfixes.)
|
1.0
|
Proposal: semver-minor and semver-patch labels - At the moment, we use the `enhancement` and `feature` labels for any semver-minor change. However, according to our [semver policy](https://github.com/eslint/eslint/blob/master/README.md#semantic-versioning-policy), not all semver-minor changes are enhancements; for example, fixing a false negative in a rule is a semver-minor change. We currently put the `enhancement` label on false-negative bugfixes to clarify that they're semver-minor, but I think this is mildly confusing; a false-negative bugfix isn't any more of an "enhancement" than a false-positive bugfix is. IMO, we're overloading the "enhancement" label to mean "semver-minor".
As a solution, I think we should add `semver-minor` and `semver-patch` labels. This would make it easier to see what can be merged for a patch release, and we would be able to use the `enhancement` label only for actual enhancements, such as new rule options.
---
I think our use `Fix:` and `Update:` in commit messages has the same problem; false-negative fixes use the `Update:` label to indicate that they're semver-minor. For example, some commit messages look like this:
```
Update: fix false negative of foo
```
But I think messages like this are easier to understand:
```
Fix: false negative of foo
```
The second commit message makes it easier for anyone skimming the changelog for new features (since they can just look at all commits with `Update:` and `New:` prefixes). At the moment, we use the `Update:` prefixes for bugfixes as well, which is a bit misleading to the reader.
(However, I know that these prefixes are used by the release tool to determine what kind of release to create, so I don't have a good solution for this problem at the moment aside from using separate prefixes for different types of bugfixes.)
|
infrastructure
|
proposal semver minor and semver patch labels at the moment we use the enhancement and feature labels for any semver minor change however according to our not all semver minor changes are enhancements for example fixing a false negative in a rule is a semver minor change we currently put the enhancement label on false negative bugfixes to clarify that they re semver minor but i think this is mildly confusing a false negative bugfix isn t any more of an enhancement than a false positive bugfix is imo we re overloading the enhancement label to mean semver minor as a solution i think we should add semver minor and semver patch labels this would make it easier to see what can be merged for a patch release and we would be able to use the enhancement label only for actual enhancements such as new rule options i think our use fix and update in commit messages has the same problem false negative fixes use the update label to indicate that they re semver minor for example some commit messages look like this update fix false negative of foo but i think messages like this are easier to understand fix false negative of foo the second commit message makes it easier for anyone skimming the changelog for new features since they can just look at all commits with update and new prefixes at the moment we use the update prefixes for bugfixes as well which is a bit misleading to the reader however i know that these prefixes are used by the release tool to determine what kind of release to create so i don t have a good solution for this problem at the moment aside from using separate prefixes for different types of bugfixes
| 1
|
23,659
| 16,503,741,625
|
IssuesEvent
|
2021-05-25 16:44:39
|
dotnet/aspnetcore
|
https://api.github.com/repos/dotnet/aspnetcore
|
closed
|
Template tests are run twice azdo + helix
|
Working area-infrastructure
|
See https://github.com/dotnet/aspnetcore/pull/30703#issuecomment-793132640 for more context, but basically today we run the template tests on both the helix job and the windows test job, this issue is to track following up and deciding what the long term plan is here. cc @dougbu
|
1.0
|
Template tests are run twice azdo + helix - See https://github.com/dotnet/aspnetcore/pull/30703#issuecomment-793132640 for more context, but basically today we run the template tests on both the helix job and the windows test job, this issue is to track following up and deciding what the long term plan is here. cc @dougbu
|
infrastructure
|
template tests are run twice azdo helix see for more context but basically today we run the template tests on both the helix job and the windows test job this issue is to track following up and deciding what the long term plan is here cc dougbu
| 1
|
4,129
| 4,828,435,690
|
IssuesEvent
|
2016-11-07 16:15:23
|
ga4gh/cgtd
|
https://api.github.com/repos/ga4gh/cgtd
|
closed
|
Simplify cgtd to be only mutable operations on ipfs
|
infrastructure question
|
Currently cgtd provides three things:
- Mutable operations on cgt/ipfs (make a submissions, add a peer)
- Convenience operations on cgt (recursively find all submissions for all peers)
- HTML GUI to cgt
I think we should simplify cgt to be purely the mutable operations. The actual content would then be directly accessed in ipfs from external applications. cgtd would listen on a specific port much like ipfs does on 5001 for mutable operations and everything else would be purely accessing ipfs directly with the knowledge of the structure of a submission (json with fields and files).
Part of my motivation is informed after working on a search engine for cgt based on elastic search. Going through cgtd to get to the data didn't seem to add any value which lead me to believe it wasn't really adding value (the read only operations that is). Also by moving to a specific port it will allow you to mutate cgtd from a machine without shelling into the actual docker while still blocking that port from external access.
For the HTML gui's those should really just be published to ipfs itself and be purely client side browsing/access of ipfs.
|
1.0
|
Simplify cgtd to be only mutable operations on ipfs - Currently cgtd provides three things:
- Mutable operations on cgt/ipfs (make a submissions, add a peer)
- Convenience operations on cgt (recursively find all submissions for all peers)
- HTML GUI to cgt
I think we should simplify cgt to be purely the mutable operations. The actual content would then be directly accessed in ipfs from external applications. cgtd would listen on a specific port much like ipfs does on 5001 for mutable operations and everything else would be purely accessing ipfs directly with the knowledge of the structure of a submission (json with fields and files).
Part of my motivation is informed after working on a search engine for cgt based on elastic search. Going through cgtd to get to the data didn't seem to add any value which lead me to believe it wasn't really adding value (the read only operations that is). Also by moving to a specific port it will allow you to mutate cgtd from a machine without shelling into the actual docker while still blocking that port from external access.
For the HTML gui's those should really just be published to ipfs itself and be purely client side browsing/access of ipfs.
|
infrastructure
|
simplify cgtd to be only mutable operations on ipfs currently cgtd provides three things mutable operations on cgt ipfs make a submissions add a peer convenience operations on cgt recursively find all submissions for all peers html gui to cgt i think we should simplify cgt to be purely the mutable operations the actual content would then be directly accessed in ipfs from external applications cgtd would listen on a specific port much like ipfs does on for mutable operations and everything else would be purely accessing ipfs directly with the knowledge of the structure of a submission json with fields and files part of my motivation is informed after working on a search engine for cgt based on elastic search going through cgtd to get to the data didn t seem to add any value which lead me to believe it wasn t really adding value the read only operations that is also by moving to a specific port it will allow you to mutate cgtd from a machine without shelling into the actual docker while still blocking that port from external access for the html gui s those should really just be published to ipfs itself and be purely client side browsing access of ipfs
| 1
|
29,406
| 23,986,692,815
|
IssuesEvent
|
2022-09-13 19:45:55
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
closed
|
buiLd runtime C:\WINDOWS was unexpected at this time.
|
area-Infrastructure-coreclr untriaged
|
Executing "F:\SourceCode\Net\runtime\src\coreclr\build-runtime.cmd" -x64 -debug -os windows
BUILD: Starting Build at 16:51:22.72
BUILD: Commencing CoreCLR product build
BUILD: Checking prerequisites
BUILD: Using CMake from E:\WorkingSoftware\VS2022\Common7\IDE\CommonExtensions\Microsoft\CMake\CMake\bin\cmake.exe
BUILD: Number of processor cores 6
BUILD: Commencing build of native components for windows.x64.Debug
BUILD: Using environment: "E:\WorkingSoftware\VS2022\VC\Auxiliary\Build\vcvarsall.bat" amd64
C:\WINDOWS was unexpected at this time.
F:\SourceCode\Net\runtime\src\coreclr\runtime.proj(61,5): error MSB3073: ๅฝไปคโ"F:\SourceCode\Net\runtime\src\coreclr\buil
d-runtime.cmd" -x64 -debug -os windowsโ exit๏ผcode 255ใ
Generate failure
F:\SourceCode\Net\runtime\src\coreclr\runtime.proj(61,5): error MSB3073: ๅฝไปคโ"F:\SourceCode\Net\runtime\src\coreclr\buil
d-runtime.cmd" -x64 -debug -os windowsโexit๏ผcode255ใ
I Using Docs \ Workflow \ Building \ CoreClr \ Document compilation This is my first compilation, I don't know why this is happening. This has been bothering me for days
Please help me. This is machine translation and also the first time I add Issue. If there is any mistake, please tell me
|
1.0
|
buiLd runtime C:\WINDOWS was unexpected at this time. -
Executing "F:\SourceCode\Net\runtime\src\coreclr\build-runtime.cmd" -x64 -debug -os windows
BUILD: Starting Build at 16:51:22.72
BUILD: Commencing CoreCLR product build
BUILD: Checking prerequisites
BUILD: Using CMake from E:\WorkingSoftware\VS2022\Common7\IDE\CommonExtensions\Microsoft\CMake\CMake\bin\cmake.exe
BUILD: Number of processor cores 6
BUILD: Commencing build of native components for windows.x64.Debug
BUILD: Using environment: "E:\WorkingSoftware\VS2022\VC\Auxiliary\Build\vcvarsall.bat" amd64
C:\WINDOWS was unexpected at this time.
F:\SourceCode\Net\runtime\src\coreclr\runtime.proj(61,5): error MSB3073: ๅฝไปคโ"F:\SourceCode\Net\runtime\src\coreclr\buil
d-runtime.cmd" -x64 -debug -os windowsโ exit๏ผcode 255ใ
Generate failure
F:\SourceCode\Net\runtime\src\coreclr\runtime.proj(61,5): error MSB3073: ๅฝไปคโ"F:\SourceCode\Net\runtime\src\coreclr\buil
d-runtime.cmd" -x64 -debug -os windowsโexit๏ผcode255ใ
I Using Docs \ Workflow \ Building \ CoreClr \ Document compilation This is my first compilation, I don't know why this is happening. This has been bothering me for days
Please help me. This is machine translation and also the first time I add Issue. If there is any mistake, please tell me
|
infrastructure
|
build runtime c windows was unexpected at this time executing f sourcecode net runtime src coreclr build runtime cmd debug os windows build starting build at build commencing coreclr product build build checking prerequisites build using cmake from e workingsoftware ide commonextensions microsoft cmake cmake bin cmake exe build number of processor cores build commencing build of native components for windows debug build using environment e workingsoftware vc auxiliary build vcvarsall bat c windows was unexpected at this time f sourcecode net runtime src coreclr runtime proj error ๅฝไปคโ f sourcecode net runtime src coreclr buil d runtime cmd debug os windowsโ exit๏ผcode ใ generate failure f sourcecode net runtime src coreclr runtime proj error ๅฝไปคโ f sourcecode net runtime src coreclr buil d runtime cmd debug os windowsโexit๏ผ ใ i using docs workflow building coreclr document compilation this is my first compilation i don t know why this is happening this has been bothering me for days please help me this is machine translation and also the first time i add issue if there is any mistake please tell me
| 1
|
85,005
| 7,959,503,759
|
IssuesEvent
|
2018-07-13 01:15:33
|
alibaba/pouch
|
https://api.github.com/repos/alibaba/pouch
|
closed
|
[help wanted] add unit-test for ListVolumes
|
areas/storage areas/test good first issue
|
### โ
. Issue Description
Add unit-test for Volume Core's ListVolumes method which locate on `storage/volume/core.go`.
### โ
ก. Describe what happened
### โ
ข. Describe what you expected to happen
### โ
ฃ. How to reproduce it (as minimally and precisely as possible)
1.
2.
3.
### โ
ค. Anything else we need to know?
please "Baiji" developer refer to this issue in your PR for the convenience for us to count your score.
### โ
ฅ. Environment:
- pouch version (use `pouch version`):
- OS (e.g. from /etc/os-release):
- Kernel (e.g. `uname -a`):
- Install tools:
- Others:
|
1.0
|
[help wanted] add unit-test for ListVolumes - ### โ
. Issue Description
Add unit-test for Volume Core's ListVolumes method which locate on `storage/volume/core.go`.
### โ
ก. Describe what happened
### โ
ข. Describe what you expected to happen
### โ
ฃ. How to reproduce it (as minimally and precisely as possible)
1.
2.
3.
### โ
ค. Anything else we need to know?
please "Baiji" developer refer to this issue in your PR for the convenience for us to count your score.
### โ
ฅ. Environment:
- pouch version (use `pouch version`):
- OS (e.g. from /etc/os-release):
- Kernel (e.g. `uname -a`):
- Install tools:
- Others:
|
non_infrastructure
|
add unit test for listvolumes โ
ฐ issue description add unit test for volume core s listvolumes method which locate on storage volume core go โ
ฑ describe what happened โ
ฒ describe what you expected to happen โ
ณ how to reproduce it as minimally and precisely as possible โ
ด anything else we need to know please baiji developer refer to this issue in your pr for the convenience for us to count your score โ
ต environment pouch version use pouch version os e g from etc os release kernel e g uname a install tools others
| 0
|
200,523
| 15,801,732,795
|
IssuesEvent
|
2021-04-03 06:20:08
|
gycc7253/ped
|
https://api.github.com/repos/gycc7253/ped
|
opened
|
reset-filter unable to clear month filter
|
severity.Medium type.DocumentationBug
|
It says in UG that 'If you currently have a filter set, such as a particular month or category', but after the reset-filter command, it still shows only records of that particular month. In this case I have entered some records of other months.
I am not sure if this is intended?

<!--session: 1617430664487-a512a64d-0c81-4bd1-9bf2-6b1bad3e3cf7-->
|
1.0
|
reset-filter unable to clear month filter - It says in UG that 'If you currently have a filter set, such as a particular month or category', but after the reset-filter command, it still shows only records of that particular month. In this case I have entered some records of other months.
I am not sure if this is intended?

<!--session: 1617430664487-a512a64d-0c81-4bd1-9bf2-6b1bad3e3cf7-->
|
non_infrastructure
|
reset filter unable to clear month filter it says in ug that if you currently have a filter set such as a particular month or category but after the reset filter command it still shows only records of that particular month in this case i have entered some records of other months i am not sure if this is intended
| 0
|
21,371
| 14,542,199,996
|
IssuesEvent
|
2020-12-15 15:26:30
|
robotology/QA
|
https://api.github.com/repos/robotology/QA
|
closed
|
installation on mac - issue
|
infrastructure software
|
Hi guys, I am re-installing everything from scratch on my Mac, following this guide:
http://wiki.icub.org/wiki/Homebrew_Mavericks
I am finding an error in this formula:
brew install `brew deps --1 yarp | sed -e ':a' -e 'N' -e '$!ba' -e 's/\n/ /gโ`
It says:
ivaldi-perso:~ arwen$ brew install `brew deps --1 yarp | sed -e ':a' -e 'N' -e '$!ba' -e 's/\n/ /g'`
Error: No available formula for yarp
This command requires a formula argument
What should I do?
|
1.0
|
installation on mac - issue - Hi guys, I am re-installing everything from scratch on my Mac, following this guide:
http://wiki.icub.org/wiki/Homebrew_Mavericks
I am finding an error in this formula:
brew install `brew deps --1 yarp | sed -e ':a' -e 'N' -e '$!ba' -e 's/\n/ /gโ`
It says:
ivaldi-perso:~ arwen$ brew install `brew deps --1 yarp | sed -e ':a' -e 'N' -e '$!ba' -e 's/\n/ /g'`
Error: No available formula for yarp
This command requires a formula argument
What should I do?
|
infrastructure
|
installation on mac issue hi guys i am re installing everything from scratch on my mac following this guide i am finding an error in this formula brew install brew deps yarp sed e a e n e ba e s n gโ it says ivaldi perso arwen brew install brew deps yarp sed e a e n e ba e s n g error no available formula for yarp this command requires a formula argument what should i do
| 1
|
75,101
| 25,532,443,121
|
IssuesEvent
|
2022-11-29 09:28:51
|
scipy/scipy
|
https://api.github.com/repos/scipy/scipy
|
closed
|
BUG: doit >=0.36 does not work with PyPy
|
defect
|
### Describe your issue.
doit uses cloudpickle, which does not work with PyPy https://github.com/cloudpipe/cloudpickle/pull/461. This is a known issue in doit https://github.com/pydoit/doit/issues/409 ~that was fixed by pinning ` cloudpickle<2.0` in version 0.34.0 - 0.35.0, but the pin was lost in 0.36.0.~ Now building SciPy can not even start on PyPy because SciPy requires doit>=0.36.
### Reproducing Code Example
```python
pypy dev.py build -j 2
```
### Error message
```shell
File "dev.py", line 126, in <module>
from doit import task_params
File "/home/runner/work/_temp/db047e5d-4a0c-4672-8fc3-540abe749108/pypy-c-jit-105829-51afa45d7c16-linux64/lib/pypy3.9/site-packages/doit/__init__.py", line 33, in <module>
from doit.doit_cmd import get_var
File "/home/runner/work/_temp/db047e5d-4a0c-4672-8fc3-540abe749108/pypy-c-jit-105829-51afa45d7c16-linux64/lib/pypy3.9/site-packages/doit/doit_cmd.py", line 16, in <module>
from .cmd_run import Run
File "/home/runner/work/_temp/db047e5d-4a0c-4672-8fc3-540abe749108/pypy-c-jit-105829-51afa45d7c16-linux64/lib/pypy3.9/site-packages/doit/cmd_run.py", line 9, in <module>
from .runner import Runner, MRunner, MThreadRunner
File "/home/runner/work/_temp/db047e5d-4a0c-4672-8fc3-540abe749108/pypy-c-jit-105829-51afa45d7c16-linux64/lib/pypy3.9/site-packages/doit/runner.py", line 8, in <module>
import cloudpickle
File "/home/runner/work/_temp/db047e5d-4a0c-4672-8fc3-540abe749108/pypy-c-jit-105829-51afa45d7c16-linux64/lib/pypy3.9/site-packages/cloudpickle/__init__.py", line 1, in <module>
from cloudpickle.cloudpickle import * # noqa
File "/home/runner/work/_temp/db047e5d-4a0c-4672-8fc3-540abe749108/pypy-c-jit-105829-51afa45d7c16-linux64/lib/pypy3.9/site-packages/cloudpickle/cloudpickle.py", line 56, in <module>
from .compat import pickle
File "/home/runner/work/_temp/db047e5d-4a0c-4672-8fc3-540abe749108/pypy-c-jit-105829-51afa45d7c16-linux64/lib/pypy3.9/site-packages/cloudpickle/compat.py", line 13, in <module>
from _pickle import Pickler # noqa: F401
ModuleNotFoundError: No module named '_pickle'
Error: Process completed with exit code 1.
```
### SciPy/NumPy/Python version information
SciPy HEAD
Edit: pinning cloudpickle does not help, there is no released version of cloudpickle that works with PyPy.
|
1.0
|
BUG: doit >=0.36 does not work with PyPy - ### Describe your issue.
doit uses cloudpickle, which does not work with PyPy https://github.com/cloudpipe/cloudpickle/pull/461. This is a known issue in doit https://github.com/pydoit/doit/issues/409 ~that was fixed by pinning ` cloudpickle<2.0` in version 0.34.0 - 0.35.0, but the pin was lost in 0.36.0.~ Now building SciPy can not even start on PyPy because SciPy requires doit>=0.36.
### Reproducing Code Example
```python
pypy dev.py build -j 2
```
### Error message
```shell
File "dev.py", line 126, in <module>
from doit import task_params
File "/home/runner/work/_temp/db047e5d-4a0c-4672-8fc3-540abe749108/pypy-c-jit-105829-51afa45d7c16-linux64/lib/pypy3.9/site-packages/doit/__init__.py", line 33, in <module>
from doit.doit_cmd import get_var
File "/home/runner/work/_temp/db047e5d-4a0c-4672-8fc3-540abe749108/pypy-c-jit-105829-51afa45d7c16-linux64/lib/pypy3.9/site-packages/doit/doit_cmd.py", line 16, in <module>
from .cmd_run import Run
File "/home/runner/work/_temp/db047e5d-4a0c-4672-8fc3-540abe749108/pypy-c-jit-105829-51afa45d7c16-linux64/lib/pypy3.9/site-packages/doit/cmd_run.py", line 9, in <module>
from .runner import Runner, MRunner, MThreadRunner
File "/home/runner/work/_temp/db047e5d-4a0c-4672-8fc3-540abe749108/pypy-c-jit-105829-51afa45d7c16-linux64/lib/pypy3.9/site-packages/doit/runner.py", line 8, in <module>
import cloudpickle
File "/home/runner/work/_temp/db047e5d-4a0c-4672-8fc3-540abe749108/pypy-c-jit-105829-51afa45d7c16-linux64/lib/pypy3.9/site-packages/cloudpickle/__init__.py", line 1, in <module>
from cloudpickle.cloudpickle import * # noqa
File "/home/runner/work/_temp/db047e5d-4a0c-4672-8fc3-540abe749108/pypy-c-jit-105829-51afa45d7c16-linux64/lib/pypy3.9/site-packages/cloudpickle/cloudpickle.py", line 56, in <module>
from .compat import pickle
File "/home/runner/work/_temp/db047e5d-4a0c-4672-8fc3-540abe749108/pypy-c-jit-105829-51afa45d7c16-linux64/lib/pypy3.9/site-packages/cloudpickle/compat.py", line 13, in <module>
from _pickle import Pickler # noqa: F401
ModuleNotFoundError: No module named '_pickle'
Error: Process completed with exit code 1.
```
### SciPy/NumPy/Python version information
SciPy HEAD
Edit: pinning cloudpickle does not help, there is no released version of cloudpickle that works with PyPy.
|
non_infrastructure
|
bug doit does not work with pypy describe your issue doit uses cloudpickle which does not work with pypy this is a known issue in doit that was fixed by pinning cloudpickle reproducing code example python pypy dev py build j error message shell file dev py line in from doit import task params file home runner work temp pypy c jit lib site packages doit init py line in from doit doit cmd import get var file home runner work temp pypy c jit lib site packages doit doit cmd py line in from cmd run import run file home runner work temp pypy c jit lib site packages doit cmd run py line in from runner import runner mrunner mthreadrunner file home runner work temp pypy c jit lib site packages doit runner py line in import cloudpickle file home runner work temp pypy c jit lib site packages cloudpickle init py line in from cloudpickle cloudpickle import noqa file home runner work temp pypy c jit lib site packages cloudpickle cloudpickle py line in from compat import pickle file home runner work temp pypy c jit lib site packages cloudpickle compat py line in from pickle import pickler noqa modulenotfounderror no module named pickle error process completed with exit code scipy numpy python version information scipy head edit pinning cloudpickle does not help there is no released version of cloudpickle that works with pypy
| 0
|
37,625
| 15,344,410,972
|
IssuesEvent
|
2021-02-28 01:02:30
|
Vivid-Project/microservice
|
https://api.github.com/repos/Vivid-Project/microservice
|
opened
|
Calc Big-O Notation
|
big-o microservice
|
Even though this microservice is light, some of the functions inside this microservice could be improved upon, and it would be beneficial in the long run to spend time ensuring that it is running as optimally as possible.
|
1.0
|
Calc Big-O Notation - Even though this microservice is light, some of the functions inside this microservice could be improved upon, and it would be beneficial in the long run to spend time ensuring that it is running as optimally as possible.
|
non_infrastructure
|
calc big o notation even though this microservice is light some of the functions inside this microservice could be improved upon and it would be beneficial in the long run to spend time ensuring that it is running as optimally as possible
| 0
|
20,248
| 2,622,336,977
|
IssuesEvent
|
2015-03-04 01:40:34
|
0xtob/nitrotracker
|
https://api.github.com/repos/0xtob/nitrotracker
|
opened
|
Add support for effects
|
auto-migrated Priority-High
|
```
Effects are missing from the GUI.
```
Original issue reported on code.google.com by `wey...@gmail.com` on 5 Apr 2010 at 5:50
|
1.0
|
Add support for effects - ```
Effects are missing from the GUI.
```
Original issue reported on code.google.com by `wey...@gmail.com` on 5 Apr 2010 at 5:50
|
non_infrastructure
|
add support for effects effects are missing from the gui original issue reported on code google com by wey gmail com on apr at
| 0
|
155,984
| 5,962,836,190
|
IssuesEvent
|
2017-05-30 01:12:15
|
input-output-hk/cardano-sl
|
https://api.github.com/repos/input-output-hk/cardano-sl
|
closed
|
[CSL-405] Check indices of headers/blocks in communication
|
Mirroring Priority:Normal State:Done Type:Task
|
<blockquote>@volhovm</blockquote>
Ensure that we send and expect similar things in terms of closed/half-closed intervals.
Seemingly bug: `handleBlocks` expects LCA to be sent from the server. It won't be sent.
<!--MIRROR_META={"service":"youtrack","id":"CSL-405"}-->
|
1.0
|
[CSL-405] Check indices of headers/blocks in communication - <blockquote>@volhovm</blockquote>
Ensure that we send and expect similar things in terms of closed/half-closed intervals.
Seemingly bug: `handleBlocks` expects LCA to be sent from the server. It won't be sent.
<!--MIRROR_META={"service":"youtrack","id":"CSL-405"}-->
|
non_infrastructure
|
check indices of headers blocks in communication volhovm ensure that we send and expect similar things in terms of closed half closed intervals seemingly bug handleblocks expects lca to be sent from the server it won t be sent
| 0
|
313,769
| 26,952,491,394
|
IssuesEvent
|
2023-02-08 12:45:32
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
roachtest: replicate/wide failed
|
C-bug C-test-failure O-robot O-roachtest branch-master release-blocker T-kv
|
roachtest.replicate/wide [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8580139?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8580139?buildTab=artifacts#/replicate/wide) on master @ [5fbcd8a8deac0205c7df38e340c1eb9692854383](https://github.com/cockroachdb/cockroach/commits/5fbcd8a8deac0205c7df38e340c1eb9692854383):
```
test artifacts and logs in: /artifacts/replicate/wide/run_1
(allocator.go:346).func1: dial tcp 34.139.136.164:26257: connect: connection refused
(test_runner.go:1110).teardownTest: test timed out (10m0s)
```
<p>Parameters: <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=1</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_fs=ext4</code>
, <code>ROACHTEST_localSSD=true</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/kv-triage
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*replicate/wide.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-24180
|
2.0
|
roachtest: replicate/wide failed - roachtest.replicate/wide [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8580139?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8580139?buildTab=artifacts#/replicate/wide) on master @ [5fbcd8a8deac0205c7df38e340c1eb9692854383](https://github.com/cockroachdb/cockroach/commits/5fbcd8a8deac0205c7df38e340c1eb9692854383):
```
test artifacts and logs in: /artifacts/replicate/wide/run_1
(allocator.go:346).func1: dial tcp 34.139.136.164:26257: connect: connection refused
(test_runner.go:1110).teardownTest: test timed out (10m0s)
```
<p>Parameters: <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=1</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_fs=ext4</code>
, <code>ROACHTEST_localSSD=true</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/kv-triage
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*replicate/wide.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-24180
|
non_infrastructure
|
roachtest replicate wide failed roachtest replicate wide with on master test artifacts and logs in artifacts replicate wide run allocator go dial tcp connect connection refused test runner go teardowntest test timed out parameters roachtest cloud gce roachtest cpu roachtest encrypted false roachtest fs roachtest localssd true roachtest ssd help see see cc cockroachdb kv triage jira issue crdb
| 0
|
103,614
| 22,351,743,169
|
IssuesEvent
|
2022-06-15 12:40:14
|
mozilla/addons-linter
|
https://api.github.com/repos/mozilla/addons-linter
|
closed
|
index.js split()
|
component: code quality priority: p5
|
As I was looking at the code, I have noticed areas that could be updated. Here are some general comments.
Example:
https://github.com/mozilla/addons-linter/blob/6d415fa41e4bead3363e330888383e4beb1f3f9f/src/dispensary/index.js#L17-L22
There should be no need to perform `split()` twice.
```js
for (const hashEntry of this._getCachedHashes()) {
const [hash, library] = hashEntry.split(' ');
this._cachedHashes[hash] = library;
}
```
Also, `.split(/\s+/)` might prove to be more robust, and error-free. e.g.
```js
for (const hashEntry of this._getCachedHashes()) {
const [hash, library] = hashEntry.split(/\s+/);
this._cachedHashes[hash] = library;
}
```
In case `forEach` is acceptable
```js
this._getCachedHashes().forEach((hashEntry) => {
const [hash, library] = hashEntry.split(/\s+/);
this._cachedHashes[hash] = library;
});
```
|
1.0
|
index.js split() - As I was looking at the code, I have noticed areas that could be updated. Here are some general comments.
Example:
https://github.com/mozilla/addons-linter/blob/6d415fa41e4bead3363e330888383e4beb1f3f9f/src/dispensary/index.js#L17-L22
There should be no need to perform `split()` twice.
```js
for (const hashEntry of this._getCachedHashes()) {
const [hash, library] = hashEntry.split(' ');
this._cachedHashes[hash] = library;
}
```
Also, `.split(/\s+/)` might prove to be more robust, and error-free. e.g.
```js
for (const hashEntry of this._getCachedHashes()) {
const [hash, library] = hashEntry.split(/\s+/);
this._cachedHashes[hash] = library;
}
```
In case `forEach` is acceptable
```js
this._getCachedHashes().forEach((hashEntry) => {
const [hash, library] = hashEntry.split(/\s+/);
this._cachedHashes[hash] = library;
});
```
|
non_infrastructure
|
index js split as i was looking at the code i have noticed areas that could be updated here are some general comments example there should be no need to perform split twice js for const hashentry of this getcachedhashes const hashentry split this cachedhashes library also split s might prove to be more robust and error free e g js for const hashentry of this getcachedhashes const hashentry split s this cachedhashes library in case foreach is acceptable js this getcachedhashes foreach hashentry const hashentry split s this cachedhashes library
| 0
|
32,598
| 6,096,998,717
|
IssuesEvent
|
2017-06-20 01:13:59
|
stan-dev/stan
|
https://api.github.com/repos/stan-dev/stan
|
opened
|
next manual, 2.16++
|
documentation
|
#### Summary:
This is the issue for suggesting fixes for the Stan manual. Please just add suggestions as comments rather than opening new issues.
#### Current Version:
v2.16.0
|
1.0
|
next manual, 2.16++ - #### Summary:
This is the issue for suggesting fixes for the Stan manual. Please just add suggestions as comments rather than opening new issues.
#### Current Version:
v2.16.0
|
non_infrastructure
|
next manual summary this is the issue for suggesting fixes for the stan manual please just add suggestions as comments rather than opening new issues current version
| 0
|
34,460
| 29,983,450,812
|
IssuesEvent
|
2023-06-25 00:18:35
|
grafana/agent
|
https://api.github.com/repos/grafana/agent
|
closed
|
Update github.com/prometheus/mysqld_exporter to v0.15.0-rc.0
|
outdated-dependency type/infrastructure
|
An update for `github.com/prometheus/mysqld_exporter` (version `v0.15.0-rc.0`) is now available. Version `v0.13.0` is currently in use.
|
1.0
|
Update github.com/prometheus/mysqld_exporter to v0.15.0-rc.0 - An update for `github.com/prometheus/mysqld_exporter` (version `v0.15.0-rc.0`) is now available. Version `v0.13.0` is currently in use.
|
infrastructure
|
update github com prometheus mysqld exporter to rc an update for github com prometheus mysqld exporter version rc is now available version is currently in use
| 1
|
33,022
| 27,151,798,444
|
IssuesEvent
|
2023-02-17 02:20:25
|
APSIMInitiative/ApsimX
|
https://api.github.com/repos/APSIMInitiative/ApsimX
|
closed
|
No display of Parameters for disabled manager scripts
|
bug interface/infrastructure
|
When selecting a disabled manager script - the code is visible when selecting the script tab, but nothing displays on the parameters tab.
|
1.0
|
No display of Parameters for disabled manager scripts - When selecting a disabled manager script - the code is visible when selecting the script tab, but nothing displays on the parameters tab.
|
infrastructure
|
no display of parameters for disabled manager scripts when selecting a disabled manager script the code is visible when selecting the script tab but nothing displays on the parameters tab
| 1
|
11,540
| 9,247,756,621
|
IssuesEvent
|
2019-03-15 02:22:24
|
APSIMInitiative/ApsimX
|
https://api.github.com/repos/APSIMInitiative/ApsimX
|
closed
|
CLEM Add basic pivot table for interacting with ResourceLedgers.
|
interface/infrastructure newfeature
|
A new feature for CLEM, designed to make pivoting data in ledgers easier.
|
1.0
|
CLEM Add basic pivot table for interacting with ResourceLedgers. - A new feature for CLEM, designed to make pivoting data in ledgers easier.
|
infrastructure
|
clem add basic pivot table for interacting with resourceledgers a new feature for clem designed to make pivoting data in ledgers easier
| 1
|
8,541
| 7,473,320,900
|
IssuesEvent
|
2018-04-03 15:03:41
|
cyberFund/cyber-search
|
https://api.github.com/repos/cyberFund/cyber-search
|
closed
|
Base Alerting Implementation For Ethereum Pump
|
Priority: High Status: Available Status: Docs Type: Infrastructure
|
- [ ] Telegram Notification Chanel
- [ ] Alerting of Stopped Indexation
- [ ] Use Kafka exporter instead of self written metrics
- [ ] Redesign Ethereum Dashboards
|
1.0
|
Base Alerting Implementation For Ethereum Pump - - [ ] Telegram Notification Chanel
- [ ] Alerting of Stopped Indexation
- [ ] Use Kafka exporter instead of self written metrics
- [ ] Redesign Ethereum Dashboards
|
infrastructure
|
base alerting implementation for ethereum pump telegram notification chanel alerting of stopped indexation use kafka exporter instead of self written metrics redesign ethereum dashboards
| 1
|
229,627
| 18,396,146,452
|
IssuesEvent
|
2021-10-12 11:36:12
|
navid-farjad/yollochat
|
https://api.github.com/repos/navid-farjad/yollochat
|
closed
|
Client online status
|
test required
|
If client is online in mini-chat, agent would see it in Requester panel in ticket page.
|
1.0
|
Client online status - If client is online in mini-chat, agent would see it in Requester panel in ticket page.
|
non_infrastructure
|
client online status if client is online in mini chat agent would see it in requester panel in ticket page
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.