Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
143,970
11,589,386,939
IssuesEvent
2020-02-24 01:56:52
climate-machine/Oceananigans.jl
https://api.github.com/repos/climate-machine/Oceananigans.jl
opened
Equatorial Rossby waves on a beta plane verification experiment
science 🌊 testing 🧪
@suyashbire1 If you still have the code you wrote up, would be good to put it on a branch or open a PR (even if half baked) so we don't forget about it and eventually merge it. Would make a great verification experiment and example.
1.0
Equatorial Rossby waves on a beta plane verification experiment - @suyashbire1 If you still have the code you wrote up, would be good to put it on a branch or open a PR (even if half baked) so we don't forget about it and eventually merge it. Would make a great verification experiment and example.
non_process
equatorial rossby waves on a beta plane verification experiment if you still have the code you wrote up would be good to put it on a branch or open a pr even if half baked so we don t forget about it and eventually merge it would make a great verification experiment and example
0
13,969
16,744,470,818
IssuesEvent
2021-06-11 13:57:24
ESMValGroup/ESMValCore
https://api.github.com/repos/ESMValGroup/ESMValCore
closed
cf-units=2.1.5 for OSX still preserves older version behaviour (<2.1.4)
preprocessor testing
`cf-units=2.1.5` installed in the esmvaltool conda env on OSX as seen from the [GA test](https://github.com/ESMValGroup/ESMValCore/runs/2759882084?check_suite_focus=true) preserves the behavior of `num2date` from an older version, <2.1.4. We need to raise this with the `cf-units` develsm @bjlittle this is a SciTools package, would you be able to look into it maybe? :beer: Results for `cf_units.num2date(time.points, time.units.name, time.units.calendar)[0].strftime('%Y%m%d%H%M')` for Linux: - Behavior for 2.1.4: " 30001161200" - Behavior for 2.1.5: "30001161200" I remember that "030001161200" result was when we were using 2.1.2 or something similar.
1.0
cf-units=2.1.5 for OSX still preserves older version behaviour (<2.1.4) - `cf-units=2.1.5` installed in the esmvaltool conda env on OSX as seen from the [GA test](https://github.com/ESMValGroup/ESMValCore/runs/2759882084?check_suite_focus=true) preserves the behavior of `num2date` from an older version, <2.1.4. We need to raise this with the `cf-units` develsm @bjlittle this is a SciTools package, would you be able to look into it maybe? :beer: Results for `cf_units.num2date(time.points, time.units.name, time.units.calendar)[0].strftime('%Y%m%d%H%M')` for Linux: - Behavior for 2.1.4: " 30001161200" - Behavior for 2.1.5: "30001161200" I remember that "030001161200" result was when we were using 2.1.2 or something similar.
process
cf units for osx still preserves older version behaviour cf units installed in the esmvaltool conda env on osx as seen from the preserves the behavior of from an older version we need to raise this with the cf units develsm bjlittle this is a scitools package would you be able to look into it maybe beer results for cf units time points time units name time units calendar strftime y m d h m for linux behavior for behavior for i remember that result was when we were using or something similar
1
195,683
22,354,030,221
IssuesEvent
2022-06-15 14:17:48
MidnightBSD/src
https://api.github.com/repos/MidnightBSD/src
closed
CVE-2022-22823 (High) detected in expatexpat-2.2.9
security vulnerability
## CVE-2022-22823 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>expatexpat-2.2.9</b></p></summary> <p> <p>Fast XML parser library in C</p> <p>Library home page: <a href=https://sourceforge.net/projects/expat/>https://sourceforge.net/projects/expat/</a></p> <p>Found in HEAD commit: <a href="https://github.com/MidnightBSD/src/commit/816463d989cc5839c1cca2efb5bf2503408507fb">816463d989cc5839c1cca2efb5bf2503408507fb</a></p> <p>Found in base branch: <b>stable/2.1</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/contrib/expat/lib/xmlparse.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> build_model in xmlparse.c in Expat (aka libexpat) before 2.4.3 has an integer overflow. <p>Publish Date: 2022-01-10 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-22823>CVE-2022-22823</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2022-22823">https://nvd.nist.gov/vuln/detail/CVE-2022-22823</a></p> <p>Release Date: 2022-01-10</p> <p>Fix Resolution: expat - 2.2.6-2+deb10u2,2.2.10-2+deb11u1,2.2.0-2+deb9u4,2.4.3-1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-22823 (High) detected in expatexpat-2.2.9 - ## CVE-2022-22823 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>expatexpat-2.2.9</b></p></summary> <p> <p>Fast XML parser library in C</p> <p>Library home page: <a href=https://sourceforge.net/projects/expat/>https://sourceforge.net/projects/expat/</a></p> <p>Found in HEAD commit: <a href="https://github.com/MidnightBSD/src/commit/816463d989cc5839c1cca2efb5bf2503408507fb">816463d989cc5839c1cca2efb5bf2503408507fb</a></p> <p>Found in base branch: <b>stable/2.1</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/contrib/expat/lib/xmlparse.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> build_model in xmlparse.c in Expat (aka libexpat) before 2.4.3 has an integer overflow. <p>Publish Date: 2022-01-10 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-22823>CVE-2022-22823</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2022-22823">https://nvd.nist.gov/vuln/detail/CVE-2022-22823</a></p> <p>Release Date: 2022-01-10</p> <p>Fix Resolution: expat - 2.2.6-2+deb10u2,2.2.10-2+deb11u1,2.2.0-2+deb9u4,2.4.3-1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in expatexpat cve high severity vulnerability vulnerable library expatexpat fast xml parser library in c library home page a href found in head commit a href found in base branch stable vulnerable source files contrib expat lib xmlparse c vulnerability details build model in xmlparse c in expat aka libexpat before has an integer overflow publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution expat step up your open source security game with mend
0
253,179
8,052,479,416
IssuesEvent
2018-08-01 19:27:04
craftercms/craftercms
https://api.github.com/repos/craftercms/craftercms
opened
[craftercms] Upgrade org.apache.commons:commons-collections4 to version 4.1
CI priority: high quality
Upgrade org.apache.commons:commons-collections4 to version 4.1 for the following modules: * Commons Utilities * Core
1.0
[craftercms] Upgrade org.apache.commons:commons-collections4 to version 4.1 - Upgrade org.apache.commons:commons-collections4 to version 4.1 for the following modules: * Commons Utilities * Core
non_process
upgrade org apache commons commons to version upgrade org apache commons commons to version for the following modules commons utilities core
0
52,771
6,654,238,047
IssuesEvent
2017-09-29 11:57:01
Microsoft/AL
https://api.github.com/repos/Microsoft/AL
closed
CaptionML for Areas compile but fail Technical validation
bydesign
Adding CaptionML to area in page compiles fine in Visual Studio Code but fails technical validation when publishing to AppSource. ![image](https://user-images.githubusercontent.com/3970209/30988679-9a7f8480-a48a-11e7-9120-fec0d3dc06db.png)
1.0
CaptionML for Areas compile but fail Technical validation - Adding CaptionML to area in page compiles fine in Visual Studio Code but fails technical validation when publishing to AppSource. ![image](https://user-images.githubusercontent.com/3970209/30988679-9a7f8480-a48a-11e7-9120-fec0d3dc06db.png)
non_process
captionml for areas compile but fail technical validation adding captionml to area in page compiles fine in visual studio code but fails technical validation when publishing to appsource
0
326,390
9,955,738,935
IssuesEvent
2019-07-05 11:59:36
ReliefApplications/bms_front
https://api.github.com/repos/ReliefApplications/bms_front
closed
UX Change - Add Beneficiary Project - Med
Medium Priority Waiting for Review visual change
Need to change this to appear as an option after clicking Create. ![image](https://user-images.githubusercontent.com/31128624/58253661-396d1900-7d69-11e9-9484-52418997bc8c.png)
1.0
UX Change - Add Beneficiary Project - Med - Need to change this to appear as an option after clicking Create. ![image](https://user-images.githubusercontent.com/31128624/58253661-396d1900-7d69-11e9-9484-52418997bc8c.png)
non_process
ux change add beneficiary project med need to change this to appear as an option after clicking create
0
21,028
27,969,931,567
IssuesEvent
2023-03-25 00:19:31
darktable-org/darktable
https://api.github.com/repos/darktable-org/darktable
closed
Mask View for Raster Mask....
feature: enhancement scope: UI scope: image processing bug: pending no-issue-activity
I know that you cannot edit a raster mask, only the mask upon which it is based and in that module nevertheless it would be nice to be able to visualize when you use it in a module. It would be nice to confirm your selected mask with visual feedback in the downstream module just in case you had more than one raster mask to choose from without having to go back to the originating module to check it before using it. As it stands now at least on my version running under windows 10 3.5.0+1749~gd9d50d302 there is no option in the module to display the raster mask in a module when you apply it?
1.0
Mask View for Raster Mask.... - I know that you cannot edit a raster mask, only the mask upon which it is based and in that module nevertheless it would be nice to be able to visualize when you use it in a module. It would be nice to confirm your selected mask with visual feedback in the downstream module just in case you had more than one raster mask to choose from without having to go back to the originating module to check it before using it. As it stands now at least on my version running under windows 10 3.5.0+1749~gd9d50d302 there is no option in the module to display the raster mask in a module when you apply it?
process
mask view for raster mask i know that you cannot edit a raster mask only the mask upon which it is based and in that module nevertheless it would be nice to be able to visualize when you use it in a module it would be nice to confirm your selected mask with visual feedback in the downstream module just in case you had more than one raster mask to choose from without having to go back to the originating module to check it before using it as it stands now at least on my version running under windows there is no option in the module to display the raster mask in a module when you apply it
1
26,444
4,710,673,233
IssuesEvent
2016-10-14 11:02:08
bridgedotnet/Bridge
https://api.github.com/repos/bridgedotnet/Bridge
closed
Bridge.fn.bind produce a memory leaks
defect
http://deck.net/cb3d8869609e9184beb85ad401227711 Bridge.fn.bind(obj, method) save reference to the "obj" into "method.$$bind" and keep it forever. This can produce huge memory leaks. Solution: ```js #before Bridge.fn.bind(obj, obj.method) #after (obj.method = Bridge.fn.bind(obj, obj.method)) ``` this will replace instance method property with a new bounded function.
1.0
Bridge.fn.bind produce a memory leaks - http://deck.net/cb3d8869609e9184beb85ad401227711 Bridge.fn.bind(obj, method) save reference to the "obj" into "method.$$bind" and keep it forever. This can produce huge memory leaks. Solution: ```js #before Bridge.fn.bind(obj, obj.method) #after (obj.method = Bridge.fn.bind(obj, obj.method)) ``` this will replace instance method property with a new bounded function.
non_process
bridge fn bind produce a memory leaks bridge fn bind obj method save reference to the obj into method bind and keep it forever this can produce huge memory leaks solution js before bridge fn bind obj obj method after obj method bridge fn bind obj obj method this will replace instance method property with a new bounded function
0
351,412
10,518,134,108
IssuesEvent
2019-09-29 08:37:03
phansch/dotfiles
https://api.github.com/repos/phansch/dotfiles
closed
Integrate ./util/dev fmt into Clippy workflow
Ansible enhancement priority:high
I don't think I want to run this on every save, but rather in a pre-commit hook.
1.0
Integrate ./util/dev fmt into Clippy workflow - I don't think I want to run this on every save, but rather in a pre-commit hook.
non_process
integrate util dev fmt into clippy workflow i don t think i want to run this on every save but rather in a pre commit hook
0
384,637
26,596,476,589
IssuesEvent
2023-01-23 12:52:56
appsmithorg/appsmith-docs
https://api.github.com/repos/appsmithorg/appsmith-docs
closed
[Docs]: Fix Inconsistent Single Quotation Marks in Bold
Documentation User Education Pod
### Is there an existing issue for this? - [X] I have searched the existing issues ### Documentation Link https://docs.appsmith.com/getting-started/start-building ### Discord/slack/intercom Link _No response_ ### Describe the problem In the [Build Your First App](https://docs.appsmith.com/getting-started/start-building) page, some of the single quotation marks are inconsistently included in the bold text. ![image](https://user-images.githubusercontent.com/112195964/207499615-f814c9e7-f8e5-42ef-9127-db945eb27766.png) ### Describe the improvement Ensure that the single quotation marks are consistently not included in the bolded text.
1.0
[Docs]: Fix Inconsistent Single Quotation Marks in Bold - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Documentation Link https://docs.appsmith.com/getting-started/start-building ### Discord/slack/intercom Link _No response_ ### Describe the problem In the [Build Your First App](https://docs.appsmith.com/getting-started/start-building) page, some of the single quotation marks are inconsistently included in the bold text. ![image](https://user-images.githubusercontent.com/112195964/207499615-f814c9e7-f8e5-42ef-9127-db945eb27766.png) ### Describe the improvement Ensure that the single quotation marks are consistently not included in the bolded text.
non_process
fix inconsistent single quotation marks in bold is there an existing issue for this i have searched the existing issues documentation link discord slack intercom link no response describe the problem in the page some of the single quotation marks are inconsistently included in the bold text describe the improvement ensure that the single quotation marks are consistently not included in the bolded text
0
20,552
27,211,080,371
IssuesEvent
2023-02-20 16:35:11
cse442-at-ub/project_s23-cinco
https://api.github.com/repos/cse442-at-ub/project_s23-cinco
closed
Create mini tutorial document for Figma
Processing Task Sprint 1
**Task Tests** *Test 1* 1) Go to https://docs.google.com/document/d/1fsVB6Wnu_IjQBG7xagPBVb3xZiERE08UYkERNkZXSgc/edit?usp=sharing 1) Assure that the information included in the document can help guide you to make a basic example page 2) Verify that all the videos included in the document have explanations of why they were included and what they are about
1.0
Create mini tutorial document for Figma - **Task Tests** *Test 1* 1) Go to https://docs.google.com/document/d/1fsVB6Wnu_IjQBG7xagPBVb3xZiERE08UYkERNkZXSgc/edit?usp=sharing 1) Assure that the information included in the document can help guide you to make a basic example page 2) Verify that all the videos included in the document have explanations of why they were included and what they are about
process
create mini tutorial document for figma task tests test go to assure that the information included in the document can help guide you to make a basic example page verify that all the videos included in the document have explanations of why they were included and what they are about
1
4,883
7,760,586,463
IssuesEvent
2018-06-01 06:46:46
Open-EO/openeo-api
https://api.github.com/repos/Open-EO/openeo-api
opened
Sharing resources
file management in discussion process graph management result access service management
It is useful to share certain resources with other users, even across back-ends. This could include process graphs, job results and web services and potentially also files. We need to think about how to incorporate sharing. Important aspects: * How to grant and revoke access. Examples: Dropbox, Google Docs, ... * How to share with individual users only (by user id?) and globally. * URLs need to be permanent (already given by using the current API URL scheme?) * What can be shared (see list above)? * Sharing files could allow users to distribute illegal content, e.g. by uploading a copyrighted movie file. * Are web services always public? * How are shared resources billed? * Does external access to web services needs to be restricted after reaching certain costs? * Does downloading job results incur a fee? * Process graphs should be always free, I assume.
1.0
Sharing resources - It is useful to share certain resources with other users, even across back-ends. This could include process graphs, job results and web services and potentially also files. We need to think about how to incorporate sharing. Important aspects: * How to grant and revoke access. Examples: Dropbox, Google Docs, ... * How to share with individual users only (by user id?) and globally. * URLs need to be permanent (already given by using the current API URL scheme?) * What can be shared (see list above)? * Sharing files could allow users to distribute illegal content, e.g. by uploading a copyrighted movie file. * Are web services always public? * How are shared resources billed? * Does external access to web services needs to be restricted after reaching certain costs? * Does downloading job results incur a fee? * Process graphs should be always free, I assume.
process
sharing resources it is useful to share certain resources with other users even across back ends this could include process graphs job results and web services and potentially also files we need to think about how to incorporate sharing important aspects how to grant and revoke access examples dropbox google docs how to share with individual users only by user id and globally urls need to be permanent already given by using the current api url scheme what can be shared see list above sharing files could allow users to distribute illegal content e g by uploading a copyrighted movie file are web services always public how are shared resources billed does external access to web services needs to be restricted after reaching certain costs does downloading job results incur a fee process graphs should be always free i assume
1
1,967
4,788,343,675
IssuesEvent
2016-10-30 14:28:10
dataproofer/Dataproofer
https://api.github.com/repos/dataproofer/Dataproofer
closed
Fix & stabilize command line, CLI, script
bug engine: processing large
I would like to use/help develop/ the CLI If I try: `node src/index.js ./sample-datasets/silk-road-arrests.csv` I'm asked about the optional suites to apply, then get various errors. If I require `fs`, and `_` at the beginning of index.js, I make a little progress, but still lots of errors: first with the filename, then another error if I tweak the filename, and so on. All of which makes me think I'm probably missing something obvious. Can you help me get on the right track? Thanks
1.0
Fix & stabilize command line, CLI, script - I would like to use/help develop/ the CLI If I try: `node src/index.js ./sample-datasets/silk-road-arrests.csv` I'm asked about the optional suites to apply, then get various errors. If I require `fs`, and `_` at the beginning of index.js, I make a little progress, but still lots of errors: first with the filename, then another error if I tweak the filename, and so on. All of which makes me think I'm probably missing something obvious. Can you help me get on the right track? Thanks
process
fix stabilize command line cli script i would like to use help develop the cli if i try node src index js sample datasets silk road arrests csv i m asked about the optional suites to apply then get various errors if i require fs and at the beginning of index js i make a little progress but still lots of errors first with the filename then another error if i tweak the filename and so on all of which makes me think i m probably missing something obvious can you help me get on the right track thanks
1
8,325
4,226,739,565
IssuesEvent
2016-07-02 17:21:13
jvanz/dwarf
https://api.github.com/repos/jvanz/dwarf
closed
Add librdkafka as submodule
build enhancement
Add the lib librdkafka as a submodule and add it in the build. https://github.com/jvanz/librdkafka PS: It's necessary enable librdkafka to be built with autotool instead of mklove (current build system)
1.0
Add librdkafka as submodule - Add the lib librdkafka as a submodule and add it in the build. https://github.com/jvanz/librdkafka PS: It's necessary enable librdkafka to be built with autotool instead of mklove (current build system)
non_process
add librdkafka as submodule add the lib librdkafka as a submodule and add it in the build ps it s necessary enable librdkafka to be built with autotool instead of mklove current build system
0
6,490
9,559,659,716
IssuesEvent
2019-05-03 17:20:51
googleapis/google-cloud-java
https://api.github.com/repos/googleapis/google-cloud-java
closed
Guava 19/20 convergence error in google-cloud-core
dependencies type: process
GAX's guava is upgraded at head but not in latest 1.30.0 release. We need a new release of GAX. We also need to upgrade protobuf-java-util to Guava 20 and push a new release. and GAX's com.google.auth:google-auth-library-oauth2-http should go to 0.11.0 ``` Dependency convergence error for com.google.guava:guava:20.0 paths to dependency are: +-com.google.cloud:zero-pom-test:0.0.1-SNAPSHOT +-com.google.cloud:google-cloud-core:1.41.0 +-com.google.guava:guava:20.0 and +-com.google.cloud:zero-pom-test:0.0.1-SNAPSHOT +-com.google.cloud:google-cloud-core:1.41.0 +-com.google.api:api-common:1.7.0 +-com.google.guava:guava:19.0 and +-com.google.cloud:zero-pom-test:0.0.1-SNAPSHOT +-com.google.cloud:google-cloud-core:1.41.0 +-com.google.api:gax:1.30.0 +-com.google.guava:guava:20.0 and +-com.google.cloud:zero-pom-test:0.0.1-SNAPSHOT +-com.google.cloud:google-cloud-core:1.41.0 +-com.google.api:gax:1.30.0 +-com.google.auth:google-auth-library-oauth2-http:0.10.0 +-com.google.guava:guava:19.0 and +-com.google.cloud:zero-pom-test:0.0.1-SNAPSHOT +-com.google.cloud:google-cloud-core:1.41.0 +-com.google.protobuf:protobuf-java-util:3.6.1 +-com.google.guava:guava:19.0 ```
1.0
Guava 19/20 convergence error in google-cloud-core - GAX's guava is upgraded at head but not in latest 1.30.0 release. We need a new release of GAX. We also need to upgrade protobuf-java-util to Guava 20 and push a new release. and GAX's com.google.auth:google-auth-library-oauth2-http should go to 0.11.0 ``` Dependency convergence error for com.google.guava:guava:20.0 paths to dependency are: +-com.google.cloud:zero-pom-test:0.0.1-SNAPSHOT +-com.google.cloud:google-cloud-core:1.41.0 +-com.google.guava:guava:20.0 and +-com.google.cloud:zero-pom-test:0.0.1-SNAPSHOT +-com.google.cloud:google-cloud-core:1.41.0 +-com.google.api:api-common:1.7.0 +-com.google.guava:guava:19.0 and +-com.google.cloud:zero-pom-test:0.0.1-SNAPSHOT +-com.google.cloud:google-cloud-core:1.41.0 +-com.google.api:gax:1.30.0 +-com.google.guava:guava:20.0 and +-com.google.cloud:zero-pom-test:0.0.1-SNAPSHOT +-com.google.cloud:google-cloud-core:1.41.0 +-com.google.api:gax:1.30.0 +-com.google.auth:google-auth-library-oauth2-http:0.10.0 +-com.google.guava:guava:19.0 and +-com.google.cloud:zero-pom-test:0.0.1-SNAPSHOT +-com.google.cloud:google-cloud-core:1.41.0 +-com.google.protobuf:protobuf-java-util:3.6.1 +-com.google.guava:guava:19.0 ```
process
guava convergence error in google cloud core gax s guava is upgraded at head but not in latest release we need a new release of gax we also need to upgrade protobuf java util to guava and push a new release and gax s com google auth google auth library http should go to dependency convergence error for com google guava guava paths to dependency are com google cloud zero pom test snapshot com google cloud google cloud core com google guava guava and com google cloud zero pom test snapshot com google cloud google cloud core com google api api common com google guava guava and com google cloud zero pom test snapshot com google cloud google cloud core com google api gax com google guava guava and com google cloud zero pom test snapshot com google cloud google cloud core com google api gax com google auth google auth library http com google guava guava and com google cloud zero pom test snapshot com google cloud google cloud core com google protobuf protobuf java util com google guava guava
1
16,999
22,364,101,814
IssuesEvent
2022-06-16 00:49:37
hashgraph/hedera-json-rpc-relay
https://api.github.com/repos/hashgraph/hedera-json-rpc-relay
opened
Update eth_getTransactionReceipt to verify return schema
enhancement P2 process
### Problem Currently the `eth_getTransactionReceipt` does not confirm the return response schema for a non existing transaction hash ### Solution Update the `eth_getTransactionReceipt` no match test to confirm the response matches ``` { "jsonrpc": "2.0", "id": 0, "result": null } ``` ### Alternatives _No response_
1.0
Update eth_getTransactionReceipt to verify return schema - ### Problem Currently the `eth_getTransactionReceipt` does not confirm the return response schema for a non existing transaction hash ### Solution Update the `eth_getTransactionReceipt` no match test to confirm the response matches ``` { "jsonrpc": "2.0", "id": 0, "result": null } ``` ### Alternatives _No response_
process
update eth gettransactionreceipt to verify return schema problem currently the eth gettransactionreceipt does not confirm the return response schema for a non existing transaction hash solution update the eth gettransactionreceipt no match test to confirm the response matches jsonrpc id result null alternatives no response
1
2,417
5,199,285,195
IssuesEvent
2017-01-23 20:27:17
elastic/beats
https://api.github.com/repos/elastic/beats
opened
Document new config options for decode_json_fields processor
:Processors docs v5.2.0
For 5.2 there are two new config options for `decode_json_fields`, namely `target` and `overwrite_keys`. Here's a description of these new fields. - `target` - Target specifies the field under which the decoded JSON shall be written. By default the decoded JSON object replaces the string field from which it was read. To merge the decoded JSON fields into the root of the event, specify target with an empty value (`target:`). - `overwrite_keys` - This specifies whether or not keys that already exist in the event are overwritten by keys from the decoded JSON object. The default value is false. Links - New options in code: https://github.com/elastic/beats/blob/5.2/libbeat/processors/actions/decode_json_fields.go#L27-L29 - Current docs: https://www.elastic.co/guide/en/beats/filebeat/master/decode-json-fields.html
1.0
Document new config options for decode_json_fields processor - For 5.2 there are two new config options for `decode_json_fields`, namely `target` and `overwrite_keys`. Here's a description of these new fields. - `target` - Target specifies the field under which the decoded JSON shall be written. By default the decoded JSON object replaces the string field from which it was read. To merge the decoded JSON fields into the root of the event, specify target with an empty value (`target:`). - `overwrite_keys` - This specifies whether or not keys that already exist in the event are overwritten by keys from the decoded JSON object. The default value is false. Links - New options in code: https://github.com/elastic/beats/blob/5.2/libbeat/processors/actions/decode_json_fields.go#L27-L29 - Current docs: https://www.elastic.co/guide/en/beats/filebeat/master/decode-json-fields.html
process
document new config options for decode json fields processor for there are two new config options for decode json fields namely target and overwrite keys here s a description of these new fields target target specifies the field under which the decoded json shall be written by default the decoded json object replaces the string field from which it was read to merge the decoded json fields into the root of the event specify target with an empty value target overwrite keys this specifies whether or not keys that already exist in the event are overwritten by keys from the decoded json object the default value is false links new options in code current docs
1
78,884
15,586,082,653
IssuesEvent
2021-03-18 01:07:57
vlaship/build-docker-image
https://api.github.com/repos/vlaship/build-docker-image
opened
CVE-2020-10968 (High) detected in jackson-databind-2.9.9.3.jar
security vulnerability
## CVE-2020-10968 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.9.3.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: build-docker-image/build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.9.3/68ddd453458765757fd3ffca9437f9a42d91003e/jackson-databind-2.9.9.3.jar,/root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.9.3/68ddd453458765757fd3ffca9437f9a42d91003e/jackson-databind-2.9.9.3.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.1.9.RELEASE.jar (Root Library) - spring-boot-starter-json-2.1.9.RELEASE.jar - jackson-module-parameter-names-2.9.9.jar - :x: **jackson-databind-2.9.9.3.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.aoju.bus.proxy.provider.remoting.RmiProvider (aka bus-proxy). <p>Publish Date: 2020-03-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10968>CVE-2020-10968</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2020-10968">https://nvd.nist.gov/vuln/detail/CVE-2020-10968</a></p> <p>Release Date: 2020-03-26</p> <p>Fix Resolution: jackson-databind-2.9.10.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-10968 (High) detected in jackson-databind-2.9.9.3.jar - ## CVE-2020-10968 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.9.3.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: build-docker-image/build.gradle</p> <p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.9.3/68ddd453458765757fd3ffca9437f9a42d91003e/jackson-databind-2.9.9.3.jar,/root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.9.3/68ddd453458765757fd3ffca9437f9a42d91003e/jackson-databind-2.9.9.3.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.1.9.RELEASE.jar (Root Library) - spring-boot-starter-json-2.1.9.RELEASE.jar - jackson-module-parameter-names-2.9.9.jar - :x: **jackson-databind-2.9.9.3.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.aoju.bus.proxy.provider.remoting.RmiProvider (aka bus-proxy). <p>Publish Date: 2020-03-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-10968>CVE-2020-10968</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2020-10968">https://nvd.nist.gov/vuln/detail/CVE-2020-10968</a></p> <p>Release Date: 2020-03-26</p> <p>Fix Resolution: jackson-databind-2.9.10.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file build docker image build gradle path to vulnerable library root gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar root gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter web release jar root library spring boot starter json release jar jackson module parameter names jar x jackson databind jar vulnerable library vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org aoju bus proxy provider remoting rmiprovider aka bus proxy publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jackson databind step up your open source security game with whitesource
0
480
2,911,386,870
IssuesEvent
2015-06-22 09:14:13
haskell-distributed/distributed-process-simplelocalnet
https://api.github.com/repos/haskell-distributed/distributed-process-simplelocalnet
opened
Add "cookie" or other identification mechanism to SimpleLocalnet
distributed-process-simplelocalnet Feature Request
_From @edsko on October 23, 2012 13:59_ so that we have multiple independent Cloud Haskell applications running on the same network. _Copied from original issue: haskell-distributed/distributed-process#56_
1.0
Add "cookie" or other identification mechanism to SimpleLocalnet - _From @edsko on October 23, 2012 13:59_ so that we have multiple independent Cloud Haskell applications running on the same network. _Copied from original issue: haskell-distributed/distributed-process#56_
process
add cookie or other identification mechanism to simplelocalnet from edsko on october so that we have multiple independent cloud haskell applications running on the same network copied from original issue haskell distributed distributed process
1
22,605
31,827,226,629
IssuesEvent
2023-09-14 08:19:44
openline-ai/openline-customer-os
https://api.github.com/repos/openline-ai/openline-customer-os
closed
[api] User should be able to remove locations from Contact
feature request app/customer-os-api app/events-processing-platform
**User should be able to remove locations:** - If location is 1-1, delete node and link - If location is 1-many, remove link
1.0
[api] User should be able to remove locations from Contact - **User should be able to remove locations:** - If location is 1-1, delete node and link - If location is 1-many, remove link
process
user should be able to remove locations from contact user should be able to remove locations if location is delete node and link if location is many remove link
1
13,117
15,504,726,461
IssuesEvent
2021-03-11 14:36:32
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Is it possible to have Scenario "Step 3" shown step by step guide?
Pri1 assigned-to-author automation/svc doc-enhancement process-automation/subsvc triaged
When reading through "Scenario 3: Start/Stop automatically based on CPU utilization" I was confused as to how to set it up. Is it possible to have that process filled with screenshots of step by step process of setting this up? When looking at the steps I seem to get lost as to what to do first from start to finish to just set up that process. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 225c9d05-83dd-b006-0025-3753f5ab25bf * Version Independent ID: 9eecef0c-b1cb-1136-faf7-542214492096 * Content: [Start/Stop VMs during off-hours solution](https://docs.microsoft.com/en-us/azure/automation/automation-solution-vm-management#feedback) * Content Source: [articles/automation/automation-solution-vm-management.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-solution-vm-management.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @bobbytreed * Microsoft Alias: **robreed**
1.0
Is it possible to have Scenario "Step 3" shown step by step guide? - When reading through "Scenario 3: Start/Stop automatically based on CPU utilization" I was confused as to how to set it up. Is it possible to have that process filled with screenshots of step by step process of setting this up? When looking at the steps I seem to get lost as to what to do first from start to finish to just set up that process. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 225c9d05-83dd-b006-0025-3753f5ab25bf * Version Independent ID: 9eecef0c-b1cb-1136-faf7-542214492096 * Content: [Start/Stop VMs during off-hours solution](https://docs.microsoft.com/en-us/azure/automation/automation-solution-vm-management#feedback) * Content Source: [articles/automation/automation-solution-vm-management.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-solution-vm-management.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @bobbytreed * Microsoft Alias: **robreed**
process
is it possible to have scenario step shown step by step guide when reading through scenario start stop automatically based on cpu utilization i was confused as to how to set it up is it possible to have that process filled with screenshots of step by step process of setting this up when looking at the steps i seem to get lost as to what to do first from start to finish to just set up that process document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login bobbytreed microsoft alias robreed
1
21,145
28,125,206,425
IssuesEvent
2023-03-31 17:06:17
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
GO:0035695 mitophagy by induced vacuole formation, definition insufficient
textual definition protein processing and quality control
I wondered precisely how this process differed from canonical autophagy (which ends with engulfment by a lysosome/vacuole) GO:0035695 mitophagy by induced vacuole formation Definition (GO:0035695 GONUTS page) The process in which cells degrade mitochondria by inducing a vacuole-like structure which directly engulfs and degrades the unhealthy mitochondria by accumulating lysosomes. PMID:21264228 and it turns out that the said 'vacuoles' are 'intramitochondrial, apparently... induces intramitochondrial lysosome-like organella that plays a critical role in the elimination of oxidized mitochondrial proteins (designated MALM for Mieap-induced accumulation of lysosome-like organelles within mitochondria). so by this definition I don'tsee how this isn't what is now described as micromitophagy? GO:0000424 micromitophagy Definition Degradation of a mitochondrion by lysosomal microautophagy. (Also GO:0035695 has only a single experimental annotation)
1.0
GO:0035695 mitophagy by induced vacuole formation, definition insufficient - I wondered precisely how this process differed from canonical autophagy (which ends with engulfment by a lysosome/vacuole) GO:0035695 mitophagy by induced vacuole formation Definition (GO:0035695 GONUTS page) The process in which cells degrade mitochondria by inducing a vacuole-like structure which directly engulfs and degrades the unhealthy mitochondria by accumulating lysosomes. PMID:21264228 and it turns out that the said 'vacuoles' are 'intramitochondrial, apparently... induces intramitochondrial lysosome-like organella that plays a critical role in the elimination of oxidized mitochondrial proteins (designated MALM for Mieap-induced accumulation of lysosome-like organelles within mitochondria). so by this definition I don'tsee how this isn't what is now described as micromitophagy? GO:0000424 micromitophagy Definition Degradation of a mitochondrion by lysosomal microautophagy. (Also GO:0035695 has only a single experimental annotation)
process
go mitophagy by induced vacuole formation definition insufficient i wondered precisely how this process differed from canonical autophagy which ends with engulfment by a lysosome vacuole go mitophagy by induced vacuole formation definition go gonuts page the process in which cells degrade mitochondria by inducing a vacuole like structure which directly engulfs and degrades the unhealthy mitochondria by accumulating lysosomes pmid and it turns out that the said vacuoles are intramitochondrial apparently induces intramitochondrial lysosome like organella that plays a critical role in the elimination of oxidized mitochondrial proteins designated malm for mieap induced accumulation of lysosome like organelles within mitochondria so by this definition i don tsee how this isn t what is now described as micromitophagy go micromitophagy definition degradation of a mitochondrion by lysosomal microautophagy also go has only a single experimental annotation
1
6,507
9,594,890,829
IssuesEvent
2019-05-09 14:54:15
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
closed
[needs-docs][processing] Rename "remove duplicates by attribute" to "delete duplicates by attribute"
Automatic new feature Easy Processing Alg
Original commit: https://github.com/qgis/QGIS/commit/d79cee1fe1108965a1e62922809ff4082dcad018 by nyalldawson for consistency with "delete duplicate geometries". Also add some tags to delete duplicate geometries algorithm.
1.0
[needs-docs][processing] Rename "remove duplicates by attribute" to "delete duplicates by attribute" - Original commit: https://github.com/qgis/QGIS/commit/d79cee1fe1108965a1e62922809ff4082dcad018 by nyalldawson for consistency with "delete duplicate geometries". Also add some tags to delete duplicate geometries algorithm.
process
rename remove duplicates by attribute to delete duplicates by attribute original commit by nyalldawson for consistency with delete duplicate geometries also add some tags to delete duplicate geometries algorithm
1
133,462
29,181,733,414
IssuesEvent
2023-05-19 12:28:55
modin-project/modin
https://api.github.com/repos/modin-project/modin
closed
REFACTOR: make series __getattr__ and __getitem__ (string) work better with non-pandas index
Code Quality 💯 P2 ponder
This is upstreaming some relevant portions of https://github.com/ponder-org/modin-public/commit/4c4dd74d95e1dc50f3940128e95fd39731078a9b context: modin-on-ponder dataframes have an index that's not stored entirely in memory, and `key in self.index` [here](https://github.com/modin-project/modin/blob/f4e96758402d757eedfb515425464eca5af25878/modin/pandas/series.py#L317) requires an extra query, while this code inside of base `__getitem__` constructs a pandas series out of the non-pandas index, which is slow because it iterates through the entire index: https://github.com/modin-project/modin/blob/f4e96758402d757eedfb515425464eca5af25878/modin/pandas/base.py#L3667-L3668 We should 1) try to getitem instead of checking the index first 1) pass a struct-like object instead of a pandas dataframe to `convert_to_index_sliceable`
1.0
REFACTOR: make series __getattr__ and __getitem__ (string) work better with non-pandas index - This is upstreaming some relevant portions of https://github.com/ponder-org/modin-public/commit/4c4dd74d95e1dc50f3940128e95fd39731078a9b context: modin-on-ponder dataframes have an index that's not stored entirely in memory, and `key in self.index` [here](https://github.com/modin-project/modin/blob/f4e96758402d757eedfb515425464eca5af25878/modin/pandas/series.py#L317) requires an extra query, while this code inside of base `__getitem__` constructs a pandas series out of the non-pandas index, which is slow because it iterates through the entire index: https://github.com/modin-project/modin/blob/f4e96758402d757eedfb515425464eca5af25878/modin/pandas/base.py#L3667-L3668 We should 1) try to getitem instead of checking the index first 1) pass a struct-like object instead of a pandas dataframe to `convert_to_index_sliceable`
non_process
refactor make series getattr and getitem string work better with non pandas index this is upstreaming some relevant portions of context modin on ponder dataframes have an index that s not stored entirely in memory and key in self index requires an extra query while this code inside of base getitem constructs a pandas series out of the non pandas index which is slow because it iterates through the entire index we should try to getitem instead of checking the index first pass a struct like object instead of a pandas dataframe to convert to index sliceable
0
3,585
3,203,793,234
IssuesEvent
2015-10-02 20:55:23
opensim-org/opensim-core
https://api.github.com/repos/opensim-org/opensim-core
closed
master branch build failing on Travis (and splitting up python wrapping)
Build
Both gcc and clang appear to be failing with internal compiler errors. That's quite a trick since they are different compilers! The build is succeeding on AppVeyor (Windows). Unless anyone is working on this I will try to reproduce it on my Ubuntu VM. /cc @aseth1 @aymanhab @chrisdembia @klshrinidhi
1.0
master branch build failing on Travis (and splitting up python wrapping) - Both gcc and clang appear to be failing with internal compiler errors. That's quite a trick since they are different compilers! The build is succeeding on AppVeyor (Windows). Unless anyone is working on this I will try to reproduce it on my Ubuntu VM. /cc @aseth1 @aymanhab @chrisdembia @klshrinidhi
non_process
master branch build failing on travis and splitting up python wrapping both gcc and clang appear to be failing with internal compiler errors that s quite a trick since they are different compilers the build is succeeding on appveyor windows unless anyone is working on this i will try to reproduce it on my ubuntu vm cc aymanhab chrisdembia klshrinidhi
0
282,624
8,708,834,006
IssuesEvent
2018-12-06 12:09:32
justin-blanco/cs1c-Group-Project
https://api.github.com/repos/justin-blanco/cs1c-Group-Project
reopened
As a user, I need to read in shapes from a saved location, the data should also be organized by a unique ID with subclasses of shape type, shape properties, and text properties.
Priority: Medium Status: Available Type: Enhancement
Description: Reads in shapes from a shapefile. Each shape in the file has an unique ID with subclasses of shape type, shape properties(, and text properties if the shape type is of type Text). Tasks: Users are able to access the saved data in the shape file. The data in the file is translated to usable data for the program Tests: Verify that the shape file is accessible Verify that the data transferred from the file to the program is usable by the program Assignee: Justin and Grant, implement functions to read in the shapes from a file into the program so that can can be displayed when called. Estimation: 15 Priority: 3 Done: Users are able to access all saved data in the shape file The data from the shapefile are properly translated, so the data can be used in the program
1.0
As a user, I need to read in shapes from a saved location, the data should also be organized by a unique ID with subclasses of shape type, shape properties, and text properties. - Description: Reads in shapes from a shapefile. Each shape in the file has an unique ID with subclasses of shape type, shape properties(, and text properties if the shape type is of type Text). Tasks: Users are able to access the saved data in the shape file. The data in the file is translated to usable data for the program Tests: Verify that the shape file is accessible Verify that the data transferred from the file to the program is usable by the program Assignee: Justin and Grant, implement functions to read in the shapes from a file into the program so that can can be displayed when called. Estimation: 15 Priority: 3 Done: Users are able to access all saved data in the shape file The data from the shapefile are properly translated, so the data can be used in the program
non_process
as a user i need to read in shapes from a saved location the data should also be organized by a unique id with subclasses of shape type shape properties and text properties description reads in shapes from a shapefile each shape in the file has an unique id with subclasses of shape type shape properties and text properties if the shape type is of type text tasks users are able to access the saved data in the shape file the data in the file is translated to usable data for the program tests verify that the shape file is accessible verify that the data transferred from the file to the program is usable by the program assignee justin and grant implement functions to read in the shapes from a file into the program so that can can be displayed when called estimation priority done users are able to access all saved data in the shape file the data from the shapefile are properly translated so the data can be used in the program
0
69,325
7,131,318,809
IssuesEvent
2018-01-22 10:33:18
SatelliteQE/robottelo
https://api.github.com/repos/SatelliteQE/robottelo
closed
UI HostContentHostUnificationTestCase - failing host rename test cases
Medium test-failure
the following test cases have been failing long-term and need some care: - test_positive_rename_content_host - test_positive_rename_content_host_cli - test_positive_rename_foreman_host
1.0
UI HostContentHostUnificationTestCase - failing host rename test cases - the following test cases have been failing long-term and need some care: - test_positive_rename_content_host - test_positive_rename_content_host_cli - test_positive_rename_foreman_host
non_process
ui hostcontenthostunificationtestcase failing host rename test cases the following test cases have been failing long term and need some care test positive rename content host test positive rename content host cli test positive rename foreman host
0
17,477
23,298,448,347
IssuesEvent
2022-08-07 00:19:25
mdsreq-fga-unb/2022.1-GDS
https://api.github.com/repos/mdsreq-fga-unb/2022.1-GDS
closed
Replanejamento do Projeto
Planejamento Processo de Desenvolvimento
**Descrição** quando será realizado o replanejamento gerenciamento dos riscos? em que momento do processo? Ou seja, quando e como a equipe irá revisar e atualizar os critérios de replanejamento do projeto?
1.0
Replanejamento do Projeto - **Descrição** quando será realizado o replanejamento gerenciamento dos riscos? em que momento do processo? Ou seja, quando e como a equipe irá revisar e atualizar os critérios de replanejamento do projeto?
process
replanejamento do projeto descrição quando será realizado o replanejamento gerenciamento dos riscos em que momento do processo ou seja quando e como a equipe irá revisar e atualizar os critérios de replanejamento do projeto
1
6,221
9,160,876,919
IssuesEvent
2019-03-01 08:56:10
fablabbcn/fablabs.io
https://api.github.com/repos/fablabbcn/fablabs.io
closed
Bug with Supernodes
Approval Process bug outdated
When [adding a new lab](https://www.fablabs.io/labs/new), any user can choose the `Supernode` category for it, but this is something than only a _superadmin_ should be able to do (the same thing applies then to the Lab edit page in the Backstage): ![screen shot 2017-05-18 at 11 32 44](https://cloud.githubusercontent.com/assets/1458229/26195974/bd851854-3bbd-11e7-85ef-627407b74959.png) Furthermore, the [Backstage](https://www.fablabs.io/backstage) should have a checkbox for searching all `Supernodes` as it is possible now for `Referees`: ![screen shot 2017-05-18 at 11 34 16](https://cloud.githubusercontent.com/assets/1458229/26196072/0c083a74-3bbe-11e7-814e-83c5023f65b9.png)
1.0
Bug with Supernodes - When [adding a new lab](https://www.fablabs.io/labs/new), any user can choose the `Supernode` category for it, but this is something than only a _superadmin_ should be able to do (the same thing applies then to the Lab edit page in the Backstage): ![screen shot 2017-05-18 at 11 32 44](https://cloud.githubusercontent.com/assets/1458229/26195974/bd851854-3bbd-11e7-85ef-627407b74959.png) Furthermore, the [Backstage](https://www.fablabs.io/backstage) should have a checkbox for searching all `Supernodes` as it is possible now for `Referees`: ![screen shot 2017-05-18 at 11 34 16](https://cloud.githubusercontent.com/assets/1458229/26196072/0c083a74-3bbe-11e7-814e-83c5023f65b9.png)
process
bug with supernodes when any user can choose the supernode category for it but this is something than only a superadmin should be able to do the same thing applies then to the lab edit page in the backstage furthermore the should have a checkbox for searching all supernodes as it is possible now for referees
1
116,042
11,898,900,349
IssuesEvent
2020-03-30 08:04:53
germanrcuriel/jira-cmd
https://api.github.com/repos/germanrcuriel/jira-cmd
closed
Document usability of api token in authentication instead of password
documentation
Document usability of api token in authentication instead of password
1.0
Document usability of api token in authentication instead of password - Document usability of api token in authentication instead of password
non_process
document usability of api token in authentication instead of password document usability of api token in authentication instead of password
0
5,478
8,355,511,244
IssuesEvent
2018-10-02 15:52:31
HumanCellAtlas/dcp-community
https://api.github.com/repos/HumanCellAtlas/dcp-community
opened
Community notifications for charters and rfcs
charter-process rfc-process
- [ ] Remove references to _This week in DCP_ from charter process - [ ] Create #dcp-announce channel for announcing charter and rfc _state changes_
2.0
Community notifications for charters and rfcs - - [ ] Remove references to _This week in DCP_ from charter process - [ ] Create #dcp-announce channel for announcing charter and rfc _state changes_
process
community notifications for charters and rfcs remove references to this week in dcp from charter process create dcp announce channel for announcing charter and rfc state changes
1
74,976
15,387,451,918
IssuesEvent
2021-03-03 09:33:02
AlexRogalskiy/github-action-screenshots
https://api.github.com/repos/AlexRogalskiy/github-action-screenshots
opened
CVE-2020-11022 (Medium) detected in jquery-1.8.1.min.js
security vulnerability
## CVE-2020-11022 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.8.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p> <p>Path to dependency file: github-action-screenshots/node_modules/redeyed/examples/browser/index.html</p> <p>Path to vulnerable library: github-action-screenshots/node_modules/redeyed/examples/browser/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.8.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/github-action-screenshots/commit/af8722bab8a58c77a11793b60fff6a42f4c2451d">af8722bab8a58c77a11793b60fff6a42f4c2451d</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jQuery - 3.5.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-11022 (Medium) detected in jquery-1.8.1.min.js - ## CVE-2020-11022 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.8.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p> <p>Path to dependency file: github-action-screenshots/node_modules/redeyed/examples/browser/index.html</p> <p>Path to vulnerable library: github-action-screenshots/node_modules/redeyed/examples/browser/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.8.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/github-action-screenshots/commit/af8722bab8a58c77a11793b60fff6a42f4c2451d">af8722bab8a58c77a11793b60fff6a42f4c2451d</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jQuery - 3.5.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file github action screenshots node modules redeyed examples browser index html path to vulnerable library github action screenshots node modules redeyed examples browser index html dependency hierarchy x jquery min js vulnerable library found in head commit a href vulnerability details in jquery versions greater than or equal to and before passing html from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource
0
211,679
16,342,437,248
IssuesEvent
2021-05-13 00:18:06
infiniteautomation/ma-core-public
https://api.github.com/repos/infiniteautomation/ma-core-public
closed
Add Edit Permission to DataPointVO - Remove cascading DataSource edit permission
For Discussion Ready for Testing
After removing the data source edit permission references from the data point: https://github.com/infiniteautomation/ma-modules-public/commit/5d41b07f5b4078155ea05e40cc762741f81051ce We are considering changing how the data source edit permission works. We would like to remove the cascading effect of this permission on data points. Currently if one has data source edit permission they can edit any of its data points. We are proposing to remove this effect and add a new permission `editPermission` to data points. During the upgrade to 4.0 we would add the data source edit roles into the new data point edit permission so the users can still edit the points.
1.0
Add Edit Permission to DataPointVO - Remove cascading DataSource edit permission - After removing the data source edit permission references from the data point: https://github.com/infiniteautomation/ma-modules-public/commit/5d41b07f5b4078155ea05e40cc762741f81051ce We are considering changing how the data source edit permission works. We would like to remove the cascading effect of this permission on data points. Currently if one has data source edit permission they can edit any of its data points. We are proposing to remove this effect and add a new permission `editPermission` to data points. During the upgrade to 4.0 we would add the data source edit roles into the new data point edit permission so the users can still edit the points.
non_process
add edit permission to datapointvo remove cascading datasource edit permission after removing the data source edit permission references from the data point we are considering changing how the data source edit permission works we would like to remove the cascading effect of this permission on data points currently if one has data source edit permission they can edit any of its data points we are proposing to remove this effect and add a new permission editpermission to data points during the upgrade to we would add the data source edit roles into the new data point edit permission so the users can still edit the points
0
6,446
9,546,272,481
IssuesEvent
2019-05-01 19:27:09
openopps/openopps-platform
https://api.github.com/repos/openopps/openopps-platform
closed
Internship email: Withdraw application confirmation
Apply Process Approved Email Requirements Ready State Dept.
Who: student What: email to let me know that I have successfully withdrawn my application Why: As an applicant I would like to be notified that I have successfully withdrawn my application A/C -The email will come from noreply@openopps.usajobs.gov  - This email will be responsive - Design teams definition of responsive : Responsive, in any context, be it email or browser, does mean responding to the viewport size (which is roughly equivalent to screen size; it's the width of the browser or email reader). These particular emails can be responsive because they include HTML, with a fallback to text for email readers that don't support HTML. -The header will be the current "Open Opportunities" branding **Subject**: U.S. Department of State Student Internship Program (Unpaid) – Application withdrawn **Trigger**: Applicant selects “Withdraw” on an opportunity they have applied for **Audience**: Applicant/Student - Insert the following content into the U.S. Department of State Student Internship Program (Unpaid) community email template #3397 **Content**: Dear [Applicant Name], You have successfully withdrawn your application with the U.S. Department of State Student Internship Program (Unpaid) for the following internships: [Insert internship opportunity 1] (Link to opportunity detail page) [Insert internship opportunity 2] (Link to opportunity detail page) [Insert internship opportunity 3] (Link to opportunity detail page) Thanks, The Open Opportunities Team
1.0
Internship email: Withdraw application confirmation - Who: student What: email to let me know that I have successfully withdrawn my application Why: As an applicant I would like to be notified that I have successfully withdrawn my application A/C -The email will come from noreply@openopps.usajobs.gov  - This email will be responsive - Design teams definition of responsive : Responsive, in any context, be it email or browser, does mean responding to the viewport size (which is roughly equivalent to screen size; it's the width of the browser or email reader). These particular emails can be responsive because they include HTML, with a fallback to text for email readers that don't support HTML. -The header will be the current "Open Opportunities" branding **Subject**: U.S. Department of State Student Internship Program (Unpaid) – Application withdrawn **Trigger**: Applicant selects “Withdraw” on an opportunity they have applied for **Audience**: Applicant/Student - Insert the following content into the U.S. Department of State Student Internship Program (Unpaid) community email template #3397 **Content**: Dear [Applicant Name], You have successfully withdrawn your application with the U.S. Department of State Student Internship Program (Unpaid) for the following internships: [Insert internship opportunity 1] (Link to opportunity detail page) [Insert internship opportunity 2] (Link to opportunity detail page) [Insert internship opportunity 3] (Link to opportunity detail page) Thanks, The Open Opportunities Team
process
internship email withdraw application confirmation who student what email to let me know that i have successfully withdrawn my application why as an applicant i would like to be notified that i have successfully withdrawn my application a c the email will come from noreply openopps usajobs gov  this email will be responsive design teams definition of responsive responsive in any context be it email or browser does mean responding to the viewport size which is roughly equivalent to screen size it s the width of the browser or email reader these particular emails can be responsive because they include html with a fallback to text for email readers that don t support html the header will be the current open opportunities branding subject u s department of state student internship program unpaid – application withdrawn trigger applicant selects “withdraw” on an opportunity they have applied for audience applicant student insert the following content into the u s department of state student internship program unpaid community email template content dear you have successfully withdrawn your application with the u s department of state student internship program unpaid for the following internships link to opportunity detail page link to opportunity detail page link to opportunity detail page thanks the open opportunities team
1
7,758
10,878,727,217
IssuesEvent
2019-11-16 19:42:16
deeplearningunb/Skincare
https://api.github.com/repos/deeplearningunb/Skincare
closed
Data cleaning
backend preprocessing
Data cleaning refers to deal with missing data and other factors that may affect badly the results of the classification model. To be accepted, the data cleaning must deal with: - [x] Missing Data - [x] Poor amount of images of a respective disease - [x] Damaged images
1.0
Data cleaning - Data cleaning refers to deal with missing data and other factors that may affect badly the results of the classification model. To be accepted, the data cleaning must deal with: - [x] Missing Data - [x] Poor amount of images of a respective disease - [x] Damaged images
process
data cleaning data cleaning refers to deal with missing data and other factors that may affect badly the results of the classification model to be accepted the data cleaning must deal with missing data poor amount of images of a respective disease damaged images
1
9,612
12,552,034,148
IssuesEvent
2020-06-06 16:44:51
dealii/dealii
https://api.github.com/repos/dealii/dealii
opened
DataOut::add_data_vector vector handling
Post-processing
The current design of `DataOut::add_data_vector` seems to be a bit annoying for several reasons: - The vectors passed to `add_data_vector` need to survive until "build_patches" - or is it the write_out functions? This makes writing modular code where the `add_data_vector` function is called in a different module than `build_patches()`/`write_...` cumbersome because one cannot use locally created vectors for output. This hit me more than once, even though I know the documentation. - One needs to make sure that the vector has ghost values updated, forcing the user to write additional code even though we would know inside of the function what kind of data we need. It would seem that this is the ideal place to also remove some of the vector templates by either equipping `add_data_vector` to copy itself to a `LinearAlgebra::read_write_vector`. Well, after the discussion at the workshop we might also only allow to pass in that kind of vectors in the first place, but that would not address the first point listed above. What do others think? @nfehn @peterrum FYI
1.0
DataOut::add_data_vector vector handling - The current design of `DataOut::add_data_vector` seems to be a bit annoying for several reasons: - The vectors passed to `add_data_vector` need to survive until "build_patches" - or is it the write_out functions? This makes writing modular code where the `add_data_vector` function is called in a different module than `build_patches()`/`write_...` cumbersome because one cannot use locally created vectors for output. This hit me more than once, even though I know the documentation. - One needs to make sure that the vector has ghost values updated, forcing the user to write additional code even though we would know inside of the function what kind of data we need. It would seem that this is the ideal place to also remove some of the vector templates by either equipping `add_data_vector` to copy itself to a `LinearAlgebra::read_write_vector`. Well, after the discussion at the workshop we might also only allow to pass in that kind of vectors in the first place, but that would not address the first point listed above. What do others think? @nfehn @peterrum FYI
process
dataout add data vector vector handling the current design of dataout add data vector seems to be a bit annoying for several reasons the vectors passed to add data vector need to survive until build patches or is it the write out functions this makes writing modular code where the add data vector function is called in a different module than build patches write cumbersome because one cannot use locally created vectors for output this hit me more than once even though i know the documentation one needs to make sure that the vector has ghost values updated forcing the user to write additional code even though we would know inside of the function what kind of data we need it would seem that this is the ideal place to also remove some of the vector templates by either equipping add data vector to copy itself to a linearalgebra read write vector well after the discussion at the workshop we might also only allow to pass in that kind of vectors in the first place but that would not address the first point listed above what do others think nfehn peterrum fyi
1
111,731
11,741,071,820
IssuesEvent
2020-03-11 20:55:57
stephenkaplan/plantpredict-python
https://api.github.com/repos/stephenkaplan/plantpredict-python
opened
There should be an even simpler and more clear setup page on documentation, potentially a video.
Documentation Example Usage
One approach could be a series of screengrab videos with audio walking through steps.
1.0
There should be an even simpler and more clear setup page on documentation, potentially a video. - One approach could be a series of screengrab videos with audio walking through steps.
non_process
there should be an even simpler and more clear setup page on documentation potentially a video one approach could be a series of screengrab videos with audio walking through steps
0
61,164
14,942,770,374
IssuesEvent
2021-01-25 21:50:27
vega/vega-lite
https://api.github.com/repos/vega/vega-lite
closed
Do not commit empty schema
Area - Refactor :building_construction: Help Wanted
When the schema generation fails, Travis creates a commit with an empty schema. We should abort in this case.
1.0
Do not commit empty schema - When the schema generation fails, Travis creates a commit with an empty schema. We should abort in this case.
non_process
do not commit empty schema when the schema generation fails travis creates a commit with an empty schema we should abort in this case
0
12,007
14,738,270,940
IssuesEvent
2021-01-07 04:16:30
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
Site Cycles - Invoice number link to report
anc-ops anc-report anp-1 ant-enhancement grt-ui processes has attachment
In GitLab by @kdjstudios on May 12, 2018, 10:46 I believe there may be an issue with how we handle the linking to reports on the Site cycle page. When clicking on the number of invoices it goes to the Invoices by site reprot. This report however is only good for finalized invoices as you can see from the screen shot below the metrics/results do not match that when we run the Review draft invoices report: ![image](/uploads/b64248db4e361e1c8c0d995cd31c13eb/image.png) [draft_invoice_detail____2_.pdf](/uploads/8cfe63fa1e23c4d7ddcf6520edeea501/draft_invoice_detail____2_.pdf) Possible solution? - We should have only completed/finalized cycle link to the invoices by site report, while non finalized should link to the draft invoices, and set the date range and site accordingly.
1.0
Site Cycles - Invoice number link to report - In GitLab by @kdjstudios on May 12, 2018, 10:46 I believe there may be an issue with how we handle the linking to reports on the Site cycle page. When clicking on the number of invoices it goes to the Invoices by site reprot. This report however is only good for finalized invoices as you can see from the screen shot below the metrics/results do not match that when we run the Review draft invoices report: ![image](/uploads/b64248db4e361e1c8c0d995cd31c13eb/image.png) [draft_invoice_detail____2_.pdf](/uploads/8cfe63fa1e23c4d7ddcf6520edeea501/draft_invoice_detail____2_.pdf) Possible solution? - We should have only completed/finalized cycle link to the invoices by site report, while non finalized should link to the draft invoices, and set the date range and site accordingly.
process
site cycles invoice number link to report in gitlab by kdjstudios on may i believe there may be an issue with how we handle the linking to reports on the site cycle page when clicking on the number of invoices it goes to the invoices by site reprot this report however is only good for finalized invoices as you can see from the screen shot below the metrics results do not match that when we run the review draft invoices report uploads image png uploads draft invoice detail pdf possible solution we should have only completed finalized cycle link to the invoices by site report while non finalized should link to the draft invoices and set the date range and site accordingly
1
139,718
18,853,751,890
IssuesEvent
2021-11-12 01:39:31
gavarasana/spring-petclinic
https://api.github.com/repos/gavarasana/spring-petclinic
opened
CVE-2021-22118 (High) detected in spring-web-5.2.8.RELEASE.jar
security vulnerability
## CVE-2021-22118 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-web-5.2.8.RELEASE.jar</b></p></summary> <p>Spring Web</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: spring-petclinic/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-web/5.2.8.RELEASE/spring-web-5.2.8.RELEASE.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.3.3.RELEASE.jar (Root Library) - :x: **spring-web-5.2.8.RELEASE.jar** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Spring Framework, versions 5.2.x prior to 5.2.15 and versions 5.3.x prior to 5.3.7, a WebFlux application is vulnerable to a privilege escalation: by (re)creating the temporary storage directory, a locally authenticated malicious user can read or modify files that have been uploaded to the WebFlux application, or overwrite arbitrary files with multipart request data. <p>Publish Date: 2021-05-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22118>CVE-2021-22118</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://tanzu.vmware.com/security/cve-2021-22118">https://tanzu.vmware.com/security/cve-2021-22118</a></p> <p>Release Date: 2021-05-27</p> <p>Fix Resolution: org.springframework:spring-web:5.2.15,5.3.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-22118 (High) detected in spring-web-5.2.8.RELEASE.jar - ## CVE-2021-22118 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-web-5.2.8.RELEASE.jar</b></p></summary> <p>Spring Web</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: spring-petclinic/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-web/5.2.8.RELEASE/spring-web-5.2.8.RELEASE.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.3.3.RELEASE.jar (Root Library) - :x: **spring-web-5.2.8.RELEASE.jar** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Spring Framework, versions 5.2.x prior to 5.2.15 and versions 5.3.x prior to 5.3.7, a WebFlux application is vulnerable to a privilege escalation: by (re)creating the temporary storage directory, a locally authenticated malicious user can read or modify files that have been uploaded to the WebFlux application, or overwrite arbitrary files with multipart request data. <p>Publish Date: 2021-05-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22118>CVE-2021-22118</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://tanzu.vmware.com/security/cve-2021-22118">https://tanzu.vmware.com/security/cve-2021-22118</a></p> <p>Release Date: 2021-05-27</p> <p>Fix Resolution: org.springframework:spring-web:5.2.15,5.3.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in spring web release jar cve high severity vulnerability vulnerable library spring web release jar spring web library home page a href path to dependency file spring petclinic pom xml path to vulnerable library home wss scanner repository org springframework spring web release spring web release jar dependency hierarchy spring boot starter web release jar root library x spring web release jar vulnerable library found in base branch main vulnerability details in spring framework versions x prior to and versions x prior to a webflux application is vulnerable to a privilege escalation by re creating the temporary storage directory a locally authenticated malicious user can read or modify files that have been uploaded to the webflux application or overwrite arbitrary files with multipart request data publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework spring web step up your open source security game with whitesource
0
159,107
13,756,036,548
IssuesEvent
2020-10-06 19:18:01
aws/sagemaker-python-sdk
https://api.github.com/repos/aws/sagemaker-python-sdk
closed
Broken link in Documentation
type: documentation
The link for [Bringing your own estimator for hyperparameter tuning](https://github.com/awslabs/amazon-sagemaker-examples/blob/master/hyperparameter_tuning/r_bring_your_own/hpo_r_bring_your_own.ipynb) in the [overview page](https://github.com/aws/sagemaker-python-sdk/blob/f99e48f93e6788281c240ff31f69eb331bd71de2/doc/overview.rst) of the documentation leads to a **page not found**· Please update the link.
1.0
Broken link in Documentation - The link for [Bringing your own estimator for hyperparameter tuning](https://github.com/awslabs/amazon-sagemaker-examples/blob/master/hyperparameter_tuning/r_bring_your_own/hpo_r_bring_your_own.ipynb) in the [overview page](https://github.com/aws/sagemaker-python-sdk/blob/f99e48f93e6788281c240ff31f69eb331bd71de2/doc/overview.rst) of the documentation leads to a **page not found**· Please update the link.
non_process
broken link in documentation the link for in the of the documentation leads to a page not found · please update the link
0
212
2,643,091,875
IssuesEvent
2015-03-12 08:07:51
arduino/Arduino
https://api.github.com/repos/arduino/Arduino
closed
Request for Hooks in the Build process
Component: Compilation Component: Preprocessor Component: Toolchain Type: Improvement
It would be very useful to be able to add a recipe to the platform specification that is executed at the start of the build process. (pre-build) In particular, my use case is to generate version strings which will be incorporated in the Rom image based on information extracted from a SCM repository. However, the concept is generic and I am sure that there are other uses. Similarly, a (post-build) hook should be provided to allow platform-specific actions at the end
1.0
Request for Hooks in the Build process - It would be very useful to be able to add a recipe to the platform specification that is executed at the start of the build process. (pre-build) In particular, my use case is to generate version strings which will be incorporated in the Rom image based on information extracted from a SCM repository. However, the concept is generic and I am sure that there are other uses. Similarly, a (post-build) hook should be provided to allow platform-specific actions at the end
process
request for hooks in the build process it would be very useful to be able to add a recipe to the platform specification that is executed at the start of the build process pre build in particular my use case is to generate version strings which will be incorporated in the rom image based on information extracted from a scm repository however the concept is generic and i am sure that there are other uses similarly a post build hook should be provided to allow platform specific actions at the end
1
22,199
30,756,016,116
IssuesEvent
2023-07-29 03:59:54
winter-telescope/mirar
https://api.github.com/repos/winter-telescope/mirar
closed
[FEATURE] De-winterify alerts
enhancement wishlist processors pipelines WINTER
**Is your feature request related to a problem? Please describe.** I'm always frustrated when I realise that the avro packet/fritz stuff is hard-coded to use a winter schema. **Describe the solution you'd like** The alert schemas should live in the pipeline directories, and the processors should take a schema as an argument.
1.0
[FEATURE] De-winterify alerts - **Is your feature request related to a problem? Please describe.** I'm always frustrated when I realise that the avro packet/fritz stuff is hard-coded to use a winter schema. **Describe the solution you'd like** The alert schemas should live in the pipeline directories, and the processors should take a schema as an argument.
process
de winterify alerts is your feature request related to a problem please describe i m always frustrated when i realise that the avro packet fritz stuff is hard coded to use a winter schema describe the solution you d like the alert schemas should live in the pipeline directories and the processors should take a schema as an argument
1
27,906
5,119,362,206
IssuesEvent
2017-01-08 17:11:26
ophrescue/RescueRails
https://api.github.com/repos/ophrescue/RescueRails
closed
No route matches [GET] "/apple-touch-icon-120x120-precomposed.png"
Defect in progress
Lots of ```No route matches [GET] "/apple-touch-icon-120x120-precomposed.png" ``` Errors all of a sudden, adding a new icon to resolve.
1.0
No route matches [GET] "/apple-touch-icon-120x120-precomposed.png" - Lots of ```No route matches [GET] "/apple-touch-icon-120x120-precomposed.png" ``` Errors all of a sudden, adding a new icon to resolve.
non_process
no route matches apple touch icon precomposed png lots of no route matches apple touch icon precomposed png errors all of a sudden adding a new icon to resolve
0
2,043
4,848,600,240
IssuesEvent
2016-11-10 17:58:24
Alfresco/alfresco-ng2-components
https://api.github.com/repos/Alfresco/alfresco-ng2-components
opened
Date does not display
browser: safari bug comp: activiti-processList
Completed tasks date displays as {{created}} and active tasks display as 'invalid date' Fine on Firefox and Chrome ![screen shot 2016-11-10 at 17 54 57](https://cloud.githubusercontent.com/assets/13200338/20188118/18f0d23e-a76f-11e6-9e70-0d0ef9938cd5.png)
1.0
Date does not display - Completed tasks date displays as {{created}} and active tasks display as 'invalid date' Fine on Firefox and Chrome ![screen shot 2016-11-10 at 17 54 57](https://cloud.githubusercontent.com/assets/13200338/20188118/18f0d23e-a76f-11e6-9e70-0d0ef9938cd5.png)
process
date does not display completed tasks date displays as created and active tasks display as invalid date fine on firefox and chrome
1
1,819
4,567,700,592
IssuesEvent
2016-09-15 12:11:26
openvstorage/openvstorage-health-check
https://api.github.com/repos/openvstorage/openvstorage-health-check
closed
Monitor restarts of all services
priority_normal process_duplicate type_feature
Apparently this already exists for arakoon, but it should be implemented for alba/voldrv/... too
1.0
Monitor restarts of all services - Apparently this already exists for arakoon, but it should be implemented for alba/voldrv/... too
process
monitor restarts of all services apparently this already exists for arakoon but it should be implemented for alba voldrv too
1
26,714
7,859,552,342
IssuesEvent
2018-06-21 16:57:55
mono/monodevelop
https://api.github.com/repos/mono/monodevelop
opened
Next/previous in structured build output should be insensitive when no results
Area: Structured Build Output vs-sync
The next/previous buttons in the structured build output should be displayed as insensitive when there are no results. The green icons look like they're sensitive. <img width="551" alt="screen shot 2018-06-21 at 12 55 22 pm" src="https://user-images.githubusercontent.com/183285/41733655-8486938c-7552-11e8-9807-cd3d38d946d8.png">
1.0
Next/previous in structured build output should be insensitive when no results - The next/previous buttons in the structured build output should be displayed as insensitive when there are no results. The green icons look like they're sensitive. <img width="551" alt="screen shot 2018-06-21 at 12 55 22 pm" src="https://user-images.githubusercontent.com/183285/41733655-8486938c-7552-11e8-9807-cd3d38d946d8.png">
non_process
next previous in structured build output should be insensitive when no results the next previous buttons in the structured build output should be displayed as insensitive when there are no results the green icons look like they re sensitive img width alt screen shot at pm src
0
22,465
31,271,843,187
IssuesEvent
2023-08-22 01:06:09
hsmusic/hsmusic-wiki
https://api.github.com/repos/hsmusic/hsmusic-wiki
opened
Automatically error for tracks that have art tags but no unique cover art
scope: data processing type: dev friendliness
Same for albums which have art tags despite not having cover art at all, I suppose, though those are even more rare. Real world example: https://github.com/hsmusic/hsmusic-data/issues/254
1.0
Automatically error for tracks that have art tags but no unique cover art - Same for albums which have art tags despite not having cover art at all, I suppose, though those are even more rare. Real world example: https://github.com/hsmusic/hsmusic-data/issues/254
process
automatically error for tracks that have art tags but no unique cover art same for albums which have art tags despite not having cover art at all i suppose though those are even more rare real world example
1
27,720
8,030,067,421
IssuesEvent
2018-07-27 18:14:47
dart-lang/build
https://api.github.com/repos/dart-lang/build
closed
Figure out what new APIs we'd need to add to support `source_gen` dynamically generating builders to wrap multiple generators
package: build_runner type: enhancement
It would be nice if packages like `source_gen` could have some hooks to dynamically create Builders which wrap multiple generators so that they can share a single output. We don't support this internally - instead we hardcode all the generators that are used. Externally this could be supported in bazel because it could happen in the 'init' step. In build_runner it could happen at the step where we generate the build step. This might not be feasible in practice, since it likely requires two phases of generating and executing scripts in order to run the package provided code.
1.0
Figure out what new APIs we'd need to add to support `source_gen` dynamically generating builders to wrap multiple generators - It would be nice if packages like `source_gen` could have some hooks to dynamically create Builders which wrap multiple generators so that they can share a single output. We don't support this internally - instead we hardcode all the generators that are used. Externally this could be supported in bazel because it could happen in the 'init' step. In build_runner it could happen at the step where we generate the build step. This might not be feasible in practice, since it likely requires two phases of generating and executing scripts in order to run the package provided code.
non_process
figure out what new apis we d need to add to support source gen dynamically generating builders to wrap multiple generators it would be nice if packages like source gen could have some hooks to dynamically create builders which wrap multiple generators so that they can share a single output we don t support this internally instead we hardcode all the generators that are used externally this could be supported in bazel because it could happen in the init step in build runner it could happen at the step where we generate the build step this might not be feasible in practice since it likely requires two phases of generating and executing scripts in order to run the package provided code
0
844
3,314,791,079
IssuesEvent
2015-11-06 08:12:25
nodejs/node
https://api.github.com/repos/nodejs/node
closed
`child_process.spawn` does not work with `npm run scripts` on windows.
child_process
`child_process.spawn` does not work with `npm run scripts` on windows. ``` # package.json { "dependencies": { "touch": "^1.0.0" }, "scripts": { "test": "npm run touch1 --verbose && node test.js", "touch1": "touch foo.txt" } } # test.js var spawn = require('child_process').spawn; var touch1 = spawn('npm', ['run', 'touch1', '--verbose'], { stdio: 'inherit' }); touch1.on('error', function(err) { console.error(err); process.exit(1); }); ``` `npm run touch1` works fine both linux and windows. `spawn('npm', ['run', 'touch1'])` works fine on linux. https://travis-ci.org/sanemat/node-windows-spawn-confirm/builds/89416274 But this does not work on windows. https://ci.appveyor.com/project/sanemat/node-windows-spawn-confirm/build/1.0.2 ``` { [Error: spawn npm ENOENT] code: 'ENOENT', errno: 'ENOENT', syscall: 'spawn npm', path: 'npm' } ``` Is this nodejs issue? or npm issue? confirming code: [child_process.spawn does not work with npm run scripts on windows. by sanemat · Pull Request #2 · sanemat/node-windows-spawn-confirm](https://github.com/sanemat/node-windows-spawn-confirm/pull/2)
1.0
`child_process.spawn` does not work with `npm run scripts` on windows. - `child_process.spawn` does not work with `npm run scripts` on windows. ``` # package.json { "dependencies": { "touch": "^1.0.0" }, "scripts": { "test": "npm run touch1 --verbose && node test.js", "touch1": "touch foo.txt" } } # test.js var spawn = require('child_process').spawn; var touch1 = spawn('npm', ['run', 'touch1', '--verbose'], { stdio: 'inherit' }); touch1.on('error', function(err) { console.error(err); process.exit(1); }); ``` `npm run touch1` works fine both linux and windows. `spawn('npm', ['run', 'touch1'])` works fine on linux. https://travis-ci.org/sanemat/node-windows-spawn-confirm/builds/89416274 But this does not work on windows. https://ci.appveyor.com/project/sanemat/node-windows-spawn-confirm/build/1.0.2 ``` { [Error: spawn npm ENOENT] code: 'ENOENT', errno: 'ENOENT', syscall: 'spawn npm', path: 'npm' } ``` Is this nodejs issue? or npm issue? confirming code: [child_process.spawn does not work with npm run scripts on windows. by sanemat · Pull Request #2 · sanemat/node-windows-spawn-confirm](https://github.com/sanemat/node-windows-spawn-confirm/pull/2)
process
child process spawn does not work with npm run scripts on windows child process spawn does not work with npm run scripts on windows package json dependencies touch scripts test npm run verbose node test js touch foo txt test js var spawn require child process spawn var spawn npm stdio inherit on error function err console error err process exit npm run works fine both linux and windows spawn npm works fine on linux but this does not work on windows code enoent errno enoent syscall spawn npm path npm is this nodejs issue or npm issue confirming code
1
16,112
20,368,013,801
IssuesEvent
2022-02-21 08:27:24
prisma/prisma
https://api.github.com/repos/prisma/prisma
closed
Unable to parse Connection String in Sharded MongoDB
bug/2-confirmed kind/bug process/candidate topic: prisma-client tech/engines team/client topic: prisma db pull topic: mongodb topic: mongodb atlas serverless
### Bug description Cannot connect successfully to mongodb when using a URL with a replica set specified. * Removing the repeated host:port portion(s) for the replica set and instead just setting a single node works fine. * The full URL with the replica set works fine when used with `mongo` on the command line. Further details: A new project using Prisma 3.9.1 (https://registry.yarnpkg.com/prisma/-/prisma-3.9.1.tgz#7510a8bf06018a5313b9427b1127ce4750b1ce5c) My mongodb connection URL, as printed from `process.env.DATABASE_URL` within my running program is as per the following (with some bits redacted, obviously). This works perfectly if I connect using `mongo` on the command line. When running in node and using `const prisma = new PrismaClient();` I get the following error: ``` (node:50866) UnhandledPromiseRejectionWarning: Error: The provided database string is invalid. Unable to parse URL. in database URL. Please refer to the documentation in [https://www.prisma.io/docs/reference/database-reference/connection-urls]() for constructing a correct connection string. In some cases, certain characters must be escaped. Please check the string for any illegal characters. at [/path/to/my/app/project/node_modules/]()@prisma[/client/runtime/index.js:36300:21]() ``` When running `npx prisma db pull --force` using the CLI I get a different error: ``` project (main*) » npx prisma db pull --force Prisma schema loaded from prisma/schema.prisma Environment variables loaded from .env Datasource "db" Introspecting based on datasource defined in prisma/schema.prisma … Error: invalid port number in database URL ``` My database URL, as printed to the console via `process.env.DATABASE_URL` immediately prior to initing the `PrismaClient` is as follows: ``` mongodb://user:password@my-shard-00.atlas.mongodb.net:27017,my-shard-01.atlas.mongodb.net:27017,my-shard-02.atlas.mongodb.net:27017/database?ssl=true&replicaSet=atlas-xxxxx-shard-0&authSource=admin&retryWrites=true ``` ### How to reproduce 1. Follow the Prisma getting started guide to create a new project using Mongo, from [here](https://www.prisma.io/docs/getting-started/setup-prisma/start-from-scratch/mongodb-typescript-mongodb). 2. Use a Mongo Atlas connection URL with a replica set 3. Cannot connect either via the prisma CLI tool or using the Prisma Client from within node. ### Expected behavior To be able to connect to a replica set. ### Prisma information From schema.prisma: ``` generator client { provider = "prisma-client-js" previewFeatures = ["mongoDb"] } datasource db { provider = "mongodb" url = env("DATABASE_URL") } ``` ### Environment & setup - OS: MacOS 12.1 - Database: Mongodb Atlas - Node.js version: v14.19.0 ### Prisma Version ``` prisma : 3.9.1 @prisma/client : 3.9.1 Current platform : darwin Query Engine (Node-API) : libquery-engine bcc2ff906db47790ee902e7bbc76d7ffb1893009 (at node_modules/@prisma/engines/libquery_engine-darwin.dylib.node) Migration Engine : migration-engine-cli bcc2ff906db47790ee902e7bbc76d7ffb1893009 (at node_modules/@prisma/engines/migration-engine-darwin) Introspection Engine : introspection-core bcc2ff906db47790ee902e7bbc76d7ffb1893009 (at node_modules/@prisma/engines/introspection-engine-darwin) Format Binary : prisma-fmt bcc2ff906db47790ee902e7bbc76d7ffb1893009 (at node_modules/@prisma/engines/prisma-fmt-darwin) Default Engines Hash : bcc2ff906db47790ee902e7bbc76d7ffb1893009 Studio : 0.457.0 Preview Features : mongoDb ```
1.0
Unable to parse Connection String in Sharded MongoDB - ### Bug description Cannot connect successfully to mongodb when using a URL with a replica set specified. * Removing the repeated host:port portion(s) for the replica set and instead just setting a single node works fine. * The full URL with the replica set works fine when used with `mongo` on the command line. Further details: A new project using Prisma 3.9.1 (https://registry.yarnpkg.com/prisma/-/prisma-3.9.1.tgz#7510a8bf06018a5313b9427b1127ce4750b1ce5c) My mongodb connection URL, as printed from `process.env.DATABASE_URL` within my running program is as per the following (with some bits redacted, obviously). This works perfectly if I connect using `mongo` on the command line. When running in node and using `const prisma = new PrismaClient();` I get the following error: ``` (node:50866) UnhandledPromiseRejectionWarning: Error: The provided database string is invalid. Unable to parse URL. in database URL. Please refer to the documentation in [https://www.prisma.io/docs/reference/database-reference/connection-urls]() for constructing a correct connection string. In some cases, certain characters must be escaped. Please check the string for any illegal characters. at [/path/to/my/app/project/node_modules/]()@prisma[/client/runtime/index.js:36300:21]() ``` When running `npx prisma db pull --force` using the CLI I get a different error: ``` project (main*) » npx prisma db pull --force Prisma schema loaded from prisma/schema.prisma Environment variables loaded from .env Datasource "db" Introspecting based on datasource defined in prisma/schema.prisma … Error: invalid port number in database URL ``` My database URL, as printed to the console via `process.env.DATABASE_URL` immediately prior to initing the `PrismaClient` is as follows: ``` mongodb://user:password@my-shard-00.atlas.mongodb.net:27017,my-shard-01.atlas.mongodb.net:27017,my-shard-02.atlas.mongodb.net:27017/database?ssl=true&replicaSet=atlas-xxxxx-shard-0&authSource=admin&retryWrites=true ``` ### How to reproduce 1. Follow the Prisma getting started guide to create a new project using Mongo, from [here](https://www.prisma.io/docs/getting-started/setup-prisma/start-from-scratch/mongodb-typescript-mongodb). 2. Use a Mongo Atlas connection URL with a replica set 3. Cannot connect either via the prisma CLI tool or using the Prisma Client from within node. ### Expected behavior To be able to connect to a replica set. ### Prisma information From schema.prisma: ``` generator client { provider = "prisma-client-js" previewFeatures = ["mongoDb"] } datasource db { provider = "mongodb" url = env("DATABASE_URL") } ``` ### Environment & setup - OS: MacOS 12.1 - Database: Mongodb Atlas - Node.js version: v14.19.0 ### Prisma Version ``` prisma : 3.9.1 @prisma/client : 3.9.1 Current platform : darwin Query Engine (Node-API) : libquery-engine bcc2ff906db47790ee902e7bbc76d7ffb1893009 (at node_modules/@prisma/engines/libquery_engine-darwin.dylib.node) Migration Engine : migration-engine-cli bcc2ff906db47790ee902e7bbc76d7ffb1893009 (at node_modules/@prisma/engines/migration-engine-darwin) Introspection Engine : introspection-core bcc2ff906db47790ee902e7bbc76d7ffb1893009 (at node_modules/@prisma/engines/introspection-engine-darwin) Format Binary : prisma-fmt bcc2ff906db47790ee902e7bbc76d7ffb1893009 (at node_modules/@prisma/engines/prisma-fmt-darwin) Default Engines Hash : bcc2ff906db47790ee902e7bbc76d7ffb1893009 Studio : 0.457.0 Preview Features : mongoDb ```
process
unable to parse connection string in sharded mongodb bug description cannot connect successfully to mongodb when using a url with a replica set specified removing the repeated host port portion s for the replica set and instead just setting a single node works fine the full url with the replica set works fine when used with mongo on the command line further details a new project using prisma my mongodb connection url as printed from process env database url within my running program is as per the following with some bits redacted obviously this works perfectly if i connect using mongo on the command line when running in node and using const prisma new prismaclient i get the following error node unhandledpromiserejectionwarning error the provided database string is invalid unable to parse url in database url please refer to the documentation in for constructing a correct connection string in some cases certain characters must be escaped please check the string for any illegal characters at prisma when running npx prisma db pull force using the cli i get a different error project main » npx prisma db pull force prisma schema loaded from prisma schema prisma environment variables loaded from env datasource db introspecting based on datasource defined in prisma schema prisma … error invalid port number in database url my database url as printed to the console via process env database url immediately prior to initing the prismaclient is as follows mongodb user password my shard atlas mongodb net my shard atlas mongodb net my shard atlas mongodb net database ssl true replicaset atlas xxxxx shard authsource admin retrywrites true how to reproduce follow the prisma getting started guide to create a new project using mongo from use a mongo atlas connection url with a replica set cannot connect either via the prisma cli tool or using the prisma client from within node expected behavior to be able to connect to a replica set prisma information from schema prisma generator client provider prisma client js previewfeatures datasource db provider mongodb url env database url environment setup os macos database mongodb atlas node js version prisma version prisma prisma client current platform darwin query engine node api libquery engine at node modules prisma engines libquery engine darwin dylib node migration engine migration engine cli at node modules prisma engines migration engine darwin introspection engine introspection core at node modules prisma engines introspection engine darwin format binary prisma fmt at node modules prisma engines prisma fmt darwin default engines hash studio preview features mongodb
1
120,774
17,644,277,610
IssuesEvent
2021-08-20 02:06:49
DavidSpek/kale
https://api.github.com/repos/DavidSpek/kale
opened
CVE-2021-29578 (High) detected in tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl, tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl
security vulnerability
## CVE-2021-29578 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl</b>, <b>tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl</b></p></summary> <p> <details><summary><b>tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/7b/c5/a97ed48fcc878e36bb05a3ea700c077360853c0994473a8f6b0ab4c2ddd2/tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/7b/c5/a97ed48fcc878e36bb05a3ea700c077360853c0994473a8f6b0ab4c2ddd2/tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl</a></p> <p>Path to dependency file: kale/examples/dog-breed-classification/requirements/requirements.txt</p> <p>Path to vulnerable library: kale/examples/dog-breed-classification/requirements/requirements.txt</p> <p> Dependency Hierarchy: - :x: **tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library) </details> <details><summary><b>tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/ef/73/205b5e7f8fe086ffe4165d984acb2c49fa3086f330f03099378753982d2e/tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/ef/73/205b5e7f8fe086ffe4165d984acb2c49fa3086f330f03099378753982d2e/tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl</a></p> <p>Path to dependency file: kale/examples/taxi-cab-classification/requirements.txt</p> <p>Path to vulnerable library: kale/examples/taxi-cab-classification/requirements.txt</p> <p> Dependency Hierarchy: - tfx_bsl-0.21.4-cp27-cp27mu-manylinux2010_x86_64.whl (Root Library) - :x: **tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl** (Vulnerable Library) </details> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> TensorFlow is an end-to-end open source platform for machine learning. The implementation of `tf.raw_ops.FractionalAvgPoolGrad` is vulnerable to a heap buffer overflow. The implementation(https://github.com/tensorflow/tensorflow/blob/dcba796a28364d6d7f003f6fe733d82726dda713/tensorflow/core/kernels/fractional_avg_pool_op.cc#L216) fails to validate that the pooling sequence arguments have enough elements as required by the `out_backprop` tensor shape. The fix will be included in TensorFlow 2.5.0. We will also cherrypick this commit on TensorFlow 2.4.2, TensorFlow 2.3.3, TensorFlow 2.2.3 and TensorFlow 2.1.4, as these are also affected and still in supported range. <p>Publish Date: 2021-05-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29578>CVE-2021-29578</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-6f89-8j54-29xf">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-6f89-8j54-29xf</a></p> <p>Release Date: 2021-05-14</p> <p>Fix Resolution: tensorflow - 2.5.0, tensorflow-cpu - 2.5.0, tensorflow-gpu - 2.5.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-29578 (High) detected in tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl, tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl - ## CVE-2021-29578 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl</b>, <b>tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl</b></p></summary> <p> <details><summary><b>tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/7b/c5/a97ed48fcc878e36bb05a3ea700c077360853c0994473a8f6b0ab4c2ddd2/tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/7b/c5/a97ed48fcc878e36bb05a3ea700c077360853c0994473a8f6b0ab4c2ddd2/tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl</a></p> <p>Path to dependency file: kale/examples/dog-breed-classification/requirements/requirements.txt</p> <p>Path to vulnerable library: kale/examples/dog-breed-classification/requirements/requirements.txt</p> <p> Dependency Hierarchy: - :x: **tensorflow-1.0.0-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library) </details> <details><summary><b>tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/ef/73/205b5e7f8fe086ffe4165d984acb2c49fa3086f330f03099378753982d2e/tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/ef/73/205b5e7f8fe086ffe4165d984acb2c49fa3086f330f03099378753982d2e/tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl</a></p> <p>Path to dependency file: kale/examples/taxi-cab-classification/requirements.txt</p> <p>Path to vulnerable library: kale/examples/taxi-cab-classification/requirements.txt</p> <p> Dependency Hierarchy: - tfx_bsl-0.21.4-cp27-cp27mu-manylinux2010_x86_64.whl (Root Library) - :x: **tensorflow-2.1.0-cp27-cp27mu-manylinux2010_x86_64.whl** (Vulnerable Library) </details> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> TensorFlow is an end-to-end open source platform for machine learning. The implementation of `tf.raw_ops.FractionalAvgPoolGrad` is vulnerable to a heap buffer overflow. The implementation(https://github.com/tensorflow/tensorflow/blob/dcba796a28364d6d7f003f6fe733d82726dda713/tensorflow/core/kernels/fractional_avg_pool_op.cc#L216) fails to validate that the pooling sequence arguments have enough elements as required by the `out_backprop` tensor shape. The fix will be included in TensorFlow 2.5.0. We will also cherrypick this commit on TensorFlow 2.4.2, TensorFlow 2.3.3, TensorFlow 2.2.3 and TensorFlow 2.1.4, as these are also affected and still in supported range. <p>Publish Date: 2021-05-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-29578>CVE-2021-29578</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-6f89-8j54-29xf">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-6f89-8j54-29xf</a></p> <p>Release Date: 2021-05-14</p> <p>Fix Resolution: tensorflow - 2.5.0, tensorflow-cpu - 2.5.0, tensorflow-gpu - 2.5.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in tensorflow whl tensorflow whl cve high severity vulnerability vulnerable libraries tensorflow whl tensorflow whl tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file kale examples dog breed classification requirements requirements txt path to vulnerable library kale examples dog breed classification requirements requirements txt dependency hierarchy x tensorflow whl vulnerable library tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file kale examples taxi cab classification requirements txt path to vulnerable library kale examples taxi cab classification requirements txt dependency hierarchy tfx bsl whl root library x tensorflow whl vulnerable library found in base branch master vulnerability details tensorflow is an end to end open source platform for machine learning the implementation of tf raw ops fractionalavgpoolgrad is vulnerable to a heap buffer overflow the implementation fails to validate that the pooling sequence arguments have enough elements as required by the out backprop tensor shape the fix will be included in tensorflow we will also cherrypick this commit on tensorflow tensorflow tensorflow and tensorflow as these are also affected and still in supported range publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow tensorflow cpu tensorflow gpu step up your open source security game with whitesource
0
134,779
30,184,818,882
IssuesEvent
2023-07-04 11:20:29
ProDemos/pds
https://api.github.com/repos/ProDemos/pds
closed
Remove support for pds-s-* scopes
documentation feature code
Has not been used - document that the actual implementation is up to designer.
1.0
Remove support for pds-s-* scopes - Has not been used - document that the actual implementation is up to designer.
non_process
remove support for pds s scopes has not been used document that the actual implementation is up to designer
0
18,980
3,114,823,593
IssuesEvent
2015-09-03 11:12:36
SciTools/iris
https://api.github.com/repos/SciTools/iris
opened
Performance of nearest vs linear
defect performance
From https://groups.google.com/forum/#!searchin/scitools-iris/performance/scitools-iris/HajdRLOZZQ4/ypy1FzbxtkMJ, with code from https://gist.github.com/pelson/8256b9bab1c775348492 investigate why the performance of nearest neighbour is ~2 orders of magnitude better than linear interpolation.
1.0
Performance of nearest vs linear - From https://groups.google.com/forum/#!searchin/scitools-iris/performance/scitools-iris/HajdRLOZZQ4/ypy1FzbxtkMJ, with code from https://gist.github.com/pelson/8256b9bab1c775348492 investigate why the performance of nearest neighbour is ~2 orders of magnitude better than linear interpolation.
non_process
performance of nearest vs linear from with code from investigate why the performance of nearest neighbour is orders of magnitude better than linear interpolation
0
16,875
22,155,359,564
IssuesEvent
2022-06-03 21:52:52
hashgraph/hedera-json-rpc-relay
https://api.github.com/repos/hashgraph/hedera-json-rpc-relay
closed
Docker compose has duplicate restart policy
bug good first issue P2 process
### Description The Docker compose file has a duplicate restart policy entry ### Steps to reproduce View file ### Additional context _No response_ ### Hedera network other ### Version v0.1.0-rc1 ### Operating system _No response_
1.0
Docker compose has duplicate restart policy - ### Description The Docker compose file has a duplicate restart policy entry ### Steps to reproduce View file ### Additional context _No response_ ### Hedera network other ### Version v0.1.0-rc1 ### Operating system _No response_
process
docker compose has duplicate restart policy description the docker compose file has a duplicate restart policy entry steps to reproduce view file additional context no response hedera network other version operating system no response
1
502
2,955,789,897
IssuesEvent
2015-07-08 06:56:29
Wordseer/wordseer
https://api.github.com/repos/Wordseer/wordseer
closed
uploading too many files puts system into broken state
Preprocessor
I tried uploading all of the files from the "personals" collection. This gave me an error which I believe originated from OS X (see attached screenshot). After that, I could no longer get the uploader tool out if its bad state (see screenshot). ![screen shot 2015-07-03 at 7 32 13 pm](https://cloud.githubusercontent.com/assets/114297/8506142/5d94b004-21ba-11e5-9acc-196fcf76dc1c.png) ![screen shot 2015-07-03 at 7 30 14 pm](https://cloud.githubusercontent.com/assets/114297/8506130/f4436d02-21b9-11e5-8120-f7b98591061c.png)
1.0
uploading too many files puts system into broken state - I tried uploading all of the files from the "personals" collection. This gave me an error which I believe originated from OS X (see attached screenshot). After that, I could no longer get the uploader tool out if its bad state (see screenshot). ![screen shot 2015-07-03 at 7 32 13 pm](https://cloud.githubusercontent.com/assets/114297/8506142/5d94b004-21ba-11e5-9acc-196fcf76dc1c.png) ![screen shot 2015-07-03 at 7 30 14 pm](https://cloud.githubusercontent.com/assets/114297/8506130/f4436d02-21b9-11e5-8120-f7b98591061c.png)
process
uploading too many files puts system into broken state i tried uploading all of the files from the personals collection this gave me an error which i believe originated from os x see attached screenshot after that i could no longer get the uploader tool out if its bad state see screenshot
1
509,567
14,739,795,315
IssuesEvent
2021-01-07 07:56:44
MikeVedsted/JoinMe
https://api.github.com/repos/MikeVedsted/JoinMe
closed
[FEAT] Add Create operation for comments on back end
Priority: Medium :zap: Status: Done :heavy_check_mark: Type: Enhancement :rocket:
**💡 I would really like to solve or include** Clearly and concisely describe the problem you are trying to solve. Currently, we do no have any create operations for comments on the backend. **👶 How would a user describe this?** Describe how users are affected by statements users might make or a user story. Why can't I comment as well? **🏆 My dream solution would be** Describe the best possible scenario of this being implemented. Route, service and handler created for comment/create **🚀 I'm ready for take off** Before submitting, please mark if you: - [x] Checked that this feature doesn't already exists - [x] Checked that a feature request doesn't already exists - [x] Went through the user flow, and understand the impact - [x] Made sure the request shows why it is important to users but doesn't exaggerate the value
1.0
[FEAT] Add Create operation for comments on back end - **💡 I would really like to solve or include** Clearly and concisely describe the problem you are trying to solve. Currently, we do no have any create operations for comments on the backend. **👶 How would a user describe this?** Describe how users are affected by statements users might make or a user story. Why can't I comment as well? **🏆 My dream solution would be** Describe the best possible scenario of this being implemented. Route, service and handler created for comment/create **🚀 I'm ready for take off** Before submitting, please mark if you: - [x] Checked that this feature doesn't already exists - [x] Checked that a feature request doesn't already exists - [x] Went through the user flow, and understand the impact - [x] Made sure the request shows why it is important to users but doesn't exaggerate the value
non_process
add create operation for comments on back end 💡 i would really like to solve or include clearly and concisely describe the problem you are trying to solve currently we do no have any create operations for comments on the backend 👶 how would a user describe this describe how users are affected by statements users might make or a user story why can t i comment as well 🏆 my dream solution would be describe the best possible scenario of this being implemented route service and handler created for comment create 🚀 i m ready for take off before submitting please mark if you checked that this feature doesn t already exists checked that a feature request doesn t already exists went through the user flow and understand the impact made sure the request shows why it is important to users but doesn t exaggerate the value
0
601,853
18,437,634,539
IssuesEvent
2021-10-14 14:34:29
encorelab/ck-board
https://api.github.com/repos/encorelab/ck-board
closed
Restrict upload background file types
bug low priority
From the board configurations (and new board creation), restrict the background image file types that can be uploaded to image file types (e.g., .jpg, .JPEG, .png, etc.) and do not allow selection of other file types (e.g., PDF, .doc, etc.). There should be an option to restrict the file types that the user can select.
1.0
Restrict upload background file types - From the board configurations (and new board creation), restrict the background image file types that can be uploaded to image file types (e.g., .jpg, .JPEG, .png, etc.) and do not allow selection of other file types (e.g., PDF, .doc, etc.). There should be an option to restrict the file types that the user can select.
non_process
restrict upload background file types from the board configurations and new board creation restrict the background image file types that can be uploaded to image file types e g jpg jpeg png etc and do not allow selection of other file types e g pdf doc etc there should be an option to restrict the file types that the user can select
0
103,090
11,339,216,703
IssuesEvent
2020-01-23 01:06:24
engnogueira/technical_interviews
https://api.github.com/repos/engnogueira/technical_interviews
opened
2.4.4 - Border Game Parte 2
documentation
Nesse tópico você vai conferir a segunda parte da resolução do exercício Border Game. Esse exercício caiu na segunda fase do processo seletivo do Google. [Border Game Parte 2](https://www.python.pro.br/modulos/entrevistas-tecnicas/topicos/border-game-parte-2) [Problema de Entrevista: Estacionamento Python](https://forum.python.pro.br/c/python) Vc deve fazer esse exercício na ordem: primeiro o problema 1, sem olhar o problema 2. Depois de terminar o problema 1, vc pode prosseguir para o segundo. Para vc ter uma ideia, essa questão rolou em um processo seletivo Europeu e era esperada a resolução de ambos passos em 30 minutos. A motivação do problema é avaliar sua habilidade com Orientação a Objetos: Problema 1: Você tem um estacionamento com 3 andares. Cada andar possui 10 vagas de carro. Construa as classes que julgar necessárias p… [Problema de OO: Balanceador de Carga (Load Balancer) Python](https://forum.python.pro.br/t/problema-de-oo-balanceador-de-carga-load-balancer/440) Segue o Enunciado: Load Balancing is quite important in Cloud environments. We are always trying to minimize the costs so we keep the number of servers as low as possible. On the other hand we know that capacity and performance improves when we add more servers. The challenge is to keep the servers as busy as possible under a certain load capacity. On our simulation environment, at each clock tick (time unit), users connect to available servers and request the same task to be executed. Each t
1.0
2.4.4 - Border Game Parte 2 - Nesse tópico você vai conferir a segunda parte da resolução do exercício Border Game. Esse exercício caiu na segunda fase do processo seletivo do Google. [Border Game Parte 2](https://www.python.pro.br/modulos/entrevistas-tecnicas/topicos/border-game-parte-2) [Problema de Entrevista: Estacionamento Python](https://forum.python.pro.br/c/python) Vc deve fazer esse exercício na ordem: primeiro o problema 1, sem olhar o problema 2. Depois de terminar o problema 1, vc pode prosseguir para o segundo. Para vc ter uma ideia, essa questão rolou em um processo seletivo Europeu e era esperada a resolução de ambos passos em 30 minutos. A motivação do problema é avaliar sua habilidade com Orientação a Objetos: Problema 1: Você tem um estacionamento com 3 andares. Cada andar possui 10 vagas de carro. Construa as classes que julgar necessárias p… [Problema de OO: Balanceador de Carga (Load Balancer) Python](https://forum.python.pro.br/t/problema-de-oo-balanceador-de-carga-load-balancer/440) Segue o Enunciado: Load Balancing is quite important in Cloud environments. We are always trying to minimize the costs so we keep the number of servers as low as possible. On the other hand we know that capacity and performance improves when we add more servers. The challenge is to keep the servers as busy as possible under a certain load capacity. On our simulation environment, at each clock tick (time unit), users connect to available servers and request the same task to be executed. Each t
non_process
border game parte nesse tópico você vai conferir a segunda parte da resolução do exercício border game esse exercício caiu na segunda fase do processo seletivo do google vc deve fazer esse exercício na ordem primeiro o problema sem olhar o problema depois de terminar o problema vc pode prosseguir para o segundo para vc ter uma ideia essa questão rolou em um processo seletivo europeu e era esperada a resolução de ambos passos em minutos a motivação do problema é avaliar sua habilidade com orientação a objetos problema você tem um estacionamento com andares cada andar possui vagas de carro construa as classes que julgar necessárias p… segue o enunciado load balancing is quite important in cloud environments we are always trying to minimize the costs so we keep the number of servers as low as possible on the other hand we know that capacity and performance improves when we add more servers the challenge is to keep the servers as busy as possible under a certain load capacity on our simulation environment at each clock tick time unit users connect to available servers and request the same task to be executed each t
0
2,493
5,268,150,805
IssuesEvent
2017-02-05 07:44:47
AllenFang/react-bootstrap-table
https://api.github.com/repos/AllenFang/react-bootstrap-table
closed
Remote Store but Local Search / Sort
inprocess
I have a remote store (using MobX) and inject the data to the table component as props. I set the `remote` flag in the table accordingly. Obviously I need to write my own `onAddRow`, `onDeleteRow` and `onCellEdit` to update the remote store, but I'd like to avoid rewriting the search and sort code. Is there a way to use the default search / sort functionality with remote store?
1.0
Remote Store but Local Search / Sort - I have a remote store (using MobX) and inject the data to the table component as props. I set the `remote` flag in the table accordingly. Obviously I need to write my own `onAddRow`, `onDeleteRow` and `onCellEdit` to update the remote store, but I'd like to avoid rewriting the search and sort code. Is there a way to use the default search / sort functionality with remote store?
process
remote store but local search sort i have a remote store using mobx and inject the data to the table component as props i set the remote flag in the table accordingly obviously i need to write my own onaddrow ondeleterow and oncelledit to update the remote store but i d like to avoid rewriting the search and sort code is there a way to use the default search sort functionality with remote store
1
9,217
12,248,992,202
IssuesEvent
2020-05-05 18:27:53
hashicorp/packer
https://api.github.com/repos/hashicorp/packer
closed
vsphere-template post-proccessor does not support vsphere-iso builder artifacts
enhancement post-processor/vsphere-template
When testing the `vsphere-template` post-processor with a `vsphere-iso` builder I get an error indicating that the post-processor can only take an artifact from the VMware-iso builder or the vsphere post-processor. Using a the `vsphere` post-processor in conjunction with the `vsphere-template` just might the be way forward. But opening the issue in case this is something we should implement. #### Reproduction Steps Run a build using the vsphere-iso builder and vsphere-template post-processor. ### Packer version v1.5.5 ### Simplified Packer Buildfile ``` { "builders": [ { "type": "vsphere-iso", "datacenter": "{{ user `datacenter_name`}}", "vcenter_server": "{{ user `vcenter_endpoint` }}", "username": "{{ user `vcenter_user` }}", "password": "{{ user `vcenter_password` }}", "host": "{{ user `esxi_host` }}", "insecure_connection": "true", "vm_name": "example-ubuntu-{{timestamp}}", "network": "VM Network", "guest_os_type": "ubuntu64Guest", "ssh_username": "vagrant", "ssh_password": "vagrant", "CPUs": 1, "RAM": 1024, "RAM_reserve_all": true, "disk_controller_type": "pvscsi", "disk_size": 32768, "disk_thin_provisioned": true, "datastore": "datastore1", "network_card": "vmxnet3", "export": { }, "iso_urls": [ "http://releases.ubuntu.com/16.04/ubuntu-16.04.6-server-amd64.iso" ], "iso_checksum": "056b7c15efc15bbbf40bf1a9ff1a3531fcbf70a2", "iso_checksum_type": "sha1", "floppy_files": [ "./preseed_hardcoded_ip.cfg" ], "boot_command": [ "<enter><wait><f6><wait><esc><wait>", "<bs><bs><bs><bs><bs><bs><bs><bs><bs><bs>", "<bs><bs><bs><bs><bs><bs><bs><bs><bs><bs>", "<bs><bs><bs><bs><bs><bs><bs><bs><bs><bs>", "<bs><bs><bs><bs><bs><bs><bs><bs><bs><bs>", "<bs><bs><bs><bs><bs><bs><bs><bs><bs><bs>", "<bs><bs><bs><bs><bs><bs><bs><bs><bs><bs>", "<bs><bs><bs><bs><bs><bs><bs><bs><bs><bs>", "<bs><bs><bs><bs><bs><bs><bs><bs><bs><bs>", "<bs><bs><bs>", "/install/vmlinuz", " initrd=/install/initrd.gz", " priority=critical", " locale=en_US", " file=/media/preseed_hardcoded_ip.cfg", " netcfg/get_ipaddress={{ user `vm_ip` }}", " netcfg/get_gateway={{ user `gateway_ip`}}", "<enter>" ] } ], "provisioners": [ { "type": "shell", "inline": ["ls /"] } ], "post-processors": [ { "type": "vsphere-template", "name": "convert-to-template-{{build_name}}", "host": "{{user `esxi_host`}}", "username": "{{user `vcenter_user`}}", "password": "{{user `vcenter_password`}}", "insecure": true, "datacenter": "{{user `datacenter_name`}}", "folder": "/pp/testing", "reregister_vm": false } ] } ``` ### Operating system and Environment details Linux, amd64 ### Log Fragments and crash.log files ``` Build 'vsphere-iso' errored: 1 error(s) occurred: * Post-processor failed: The Packer vSphere Template post-processor can only take an artifact from the VMware-iso builder, built on ESXi (i.e. remote) or an artifact from the vSphere post-processor. Artifact type jetbrains.vsphere does not fit this requirement ==> Some builds didn't complete successfully and had errors: --> vsphere-iso: 1 error(s) occurred: * Post-processor failed: The Packer vSphere Template post-processor can only take an artifact from the VMware-iso builder, built on ESXi (i.e. remote) or an artifact from the vSphere post-processor. Artifact type jetbrains.vsphere does not fit this requirement ```
1.0
vsphere-template post-proccessor does not support vsphere-iso builder artifacts - When testing the `vsphere-template` post-processor with a `vsphere-iso` builder I get an error indicating that the post-processor can only take an artifact from the VMware-iso builder or the vsphere post-processor. Using a the `vsphere` post-processor in conjunction with the `vsphere-template` just might the be way forward. But opening the issue in case this is something we should implement. #### Reproduction Steps Run a build using the vsphere-iso builder and vsphere-template post-processor. ### Packer version v1.5.5 ### Simplified Packer Buildfile ``` { "builders": [ { "type": "vsphere-iso", "datacenter": "{{ user `datacenter_name`}}", "vcenter_server": "{{ user `vcenter_endpoint` }}", "username": "{{ user `vcenter_user` }}", "password": "{{ user `vcenter_password` }}", "host": "{{ user `esxi_host` }}", "insecure_connection": "true", "vm_name": "example-ubuntu-{{timestamp}}", "network": "VM Network", "guest_os_type": "ubuntu64Guest", "ssh_username": "vagrant", "ssh_password": "vagrant", "CPUs": 1, "RAM": 1024, "RAM_reserve_all": true, "disk_controller_type": "pvscsi", "disk_size": 32768, "disk_thin_provisioned": true, "datastore": "datastore1", "network_card": "vmxnet3", "export": { }, "iso_urls": [ "http://releases.ubuntu.com/16.04/ubuntu-16.04.6-server-amd64.iso" ], "iso_checksum": "056b7c15efc15bbbf40bf1a9ff1a3531fcbf70a2", "iso_checksum_type": "sha1", "floppy_files": [ "./preseed_hardcoded_ip.cfg" ], "boot_command": [ "<enter><wait><f6><wait><esc><wait>", "<bs><bs><bs><bs><bs><bs><bs><bs><bs><bs>", "<bs><bs><bs><bs><bs><bs><bs><bs><bs><bs>", "<bs><bs><bs><bs><bs><bs><bs><bs><bs><bs>", "<bs><bs><bs><bs><bs><bs><bs><bs><bs><bs>", "<bs><bs><bs><bs><bs><bs><bs><bs><bs><bs>", "<bs><bs><bs><bs><bs><bs><bs><bs><bs><bs>", "<bs><bs><bs><bs><bs><bs><bs><bs><bs><bs>", "<bs><bs><bs><bs><bs><bs><bs><bs><bs><bs>", "<bs><bs><bs>", "/install/vmlinuz", " initrd=/install/initrd.gz", " priority=critical", " locale=en_US", " file=/media/preseed_hardcoded_ip.cfg", " netcfg/get_ipaddress={{ user `vm_ip` }}", " netcfg/get_gateway={{ user `gateway_ip`}}", "<enter>" ] } ], "provisioners": [ { "type": "shell", "inline": ["ls /"] } ], "post-processors": [ { "type": "vsphere-template", "name": "convert-to-template-{{build_name}}", "host": "{{user `esxi_host`}}", "username": "{{user `vcenter_user`}}", "password": "{{user `vcenter_password`}}", "insecure": true, "datacenter": "{{user `datacenter_name`}}", "folder": "/pp/testing", "reregister_vm": false } ] } ``` ### Operating system and Environment details Linux, amd64 ### Log Fragments and crash.log files ``` Build 'vsphere-iso' errored: 1 error(s) occurred: * Post-processor failed: The Packer vSphere Template post-processor can only take an artifact from the VMware-iso builder, built on ESXi (i.e. remote) or an artifact from the vSphere post-processor. Artifact type jetbrains.vsphere does not fit this requirement ==> Some builds didn't complete successfully and had errors: --> vsphere-iso: 1 error(s) occurred: * Post-processor failed: The Packer vSphere Template post-processor can only take an artifact from the VMware-iso builder, built on ESXi (i.e. remote) or an artifact from the vSphere post-processor. Artifact type jetbrains.vsphere does not fit this requirement ```
process
vsphere template post proccessor does not support vsphere iso builder artifacts when testing the vsphere template post processor with a vsphere iso builder i get an error indicating that the post processor can only take an artifact from the vmware iso builder or the vsphere post processor using a the vsphere post processor in conjunction with the vsphere template just might the be way forward but opening the issue in case this is something we should implement reproduction steps run a build using the vsphere iso builder and vsphere template post processor packer version simplified packer buildfile builders type vsphere iso datacenter user datacenter name vcenter server user vcenter endpoint username user vcenter user password user vcenter password host user esxi host insecure connection true vm name example ubuntu timestamp network vm network guest os type ssh username vagrant ssh password vagrant cpus ram ram reserve all true disk controller type pvscsi disk size disk thin provisioned true datastore network card export iso urls iso checksum iso checksum type floppy files preseed hardcoded ip cfg boot command install vmlinuz initrd install initrd gz priority critical locale en us file media preseed hardcoded ip cfg netcfg get ipaddress user vm ip netcfg get gateway user gateway ip provisioners type shell inline post processors type vsphere template name convert to template build name host user esxi host username user vcenter user password user vcenter password insecure true datacenter user datacenter name folder pp testing reregister vm false operating system and environment details linux log fragments and crash log files build vsphere iso errored error s occurred post processor failed the packer vsphere template post processor can only take an artifact from the vmware iso builder built on esxi i e remote or an artifact from the vsphere post processor artifact type jetbrains vsphere does not fit this requirement some builds didn t complete successfully and had errors vsphere iso error s occurred post processor failed the packer vsphere template post processor can only take an artifact from the vmware iso builder built on esxi i e remote or an artifact from the vsphere post processor artifact type jetbrains vsphere does not fit this requirement
1
118,347
4,737,760,987
IssuesEvent
2016-10-20 00:12:18
a8cteam51/smittenkitchen
https://api.github.com/repos/a8cteam51/smittenkitchen
opened
Google AMP Pages Not Showing Template, Comments or Ads
high-priority
Google, as you probably know, has introduced AMP web results for mobile and the from SK look all wrong. https://smittenkitchen.com/2016/10/winter-squash-pancakes-with-crispy-sage-and-brown-butter/amp/ Less urgent than the design is that ads and comments are missing. More than 50% of site traffic comes from mobile and we're not showing ads to these people. And readers are upset the comments have disappeared. Can we put this first in the queue? This is a pretty big problem. Thank you.
1.0
Google AMP Pages Not Showing Template, Comments or Ads - Google, as you probably know, has introduced AMP web results for mobile and the from SK look all wrong. https://smittenkitchen.com/2016/10/winter-squash-pancakes-with-crispy-sage-and-brown-butter/amp/ Less urgent than the design is that ads and comments are missing. More than 50% of site traffic comes from mobile and we're not showing ads to these people. And readers are upset the comments have disappeared. Can we put this first in the queue? This is a pretty big problem. Thank you.
non_process
google amp pages not showing template comments or ads google as you probably know has introduced amp web results for mobile and the from sk look all wrong less urgent than the design is that ads and comments are missing more than of site traffic comes from mobile and we re not showing ads to these people and readers are upset the comments have disappeared can we put this first in the queue this is a pretty big problem thank you
0
795,783
28,086,116,331
IssuesEvent
2023-03-30 09:52:59
NikkelM/Random-YouTube-Video
https://api.github.com/repos/NikkelM/Random-YouTube-Video
closed
[Bug] Catch case where all videos were deleted/no videos are uploaded on channel
Bug Priority: High
Currently, we get into an endless loop.
1.0
[Bug] Catch case where all videos were deleted/no videos are uploaded on channel - Currently, we get into an endless loop.
non_process
catch case where all videos were deleted no videos are uploaded on channel currently we get into an endless loop
0
288,869
24,942,277,745
IssuesEvent
2022-10-31 20:02:18
vmware-tanzu/community-edition
https://api.github.com/repos/vmware-tanzu/community-edition
closed
Verify management cluster e2e on AWS is working correctly
kind/test-release owner/release-eng
## Verify e2e for management-clusters on AWS This issue tracks work for AWS E2E tests for management clusters. ### Verification - [ ] Execute/run e2e tests for management clusters. This should exercise: - [ ] create cluster - [ ] simple inspection of the cluster to make sure it's running - [ ] delete cluster - [ ] simple inspection to make sure the cluster is deleted - [ ] Is this workflow being triggered by the daily build? - [ ] Is this workflow being triggered by modifications to Makefile (for a Tanzu Framework version bump)? ### Issues found as a result of the audit? We should prioritize these and determine if they are absolutely required for the workflow to function correctly and be reproducible. - [ ] [Run AWS and Azure E2E tests for daily official builds](https://github.com/vmware-tanzu/community-edition/issues/3485) - [ ] [Fix AWS E2E test pipeline failure: NAT Gateway issue](https://github.com/vmware-tanzu/community-edition/issues/2739) - [ ] [Run AWS E2E tests for Tanzu Framework version bump PRs](https://github.com/vmware-tanzu/community-edition/issues/1519)
1.0
Verify management cluster e2e on AWS is working correctly - ## Verify e2e for management-clusters on AWS This issue tracks work for AWS E2E tests for management clusters. ### Verification - [ ] Execute/run e2e tests for management clusters. This should exercise: - [ ] create cluster - [ ] simple inspection of the cluster to make sure it's running - [ ] delete cluster - [ ] simple inspection to make sure the cluster is deleted - [ ] Is this workflow being triggered by the daily build? - [ ] Is this workflow being triggered by modifications to Makefile (for a Tanzu Framework version bump)? ### Issues found as a result of the audit? We should prioritize these and determine if they are absolutely required for the workflow to function correctly and be reproducible. - [ ] [Run AWS and Azure E2E tests for daily official builds](https://github.com/vmware-tanzu/community-edition/issues/3485) - [ ] [Fix AWS E2E test pipeline failure: NAT Gateway issue](https://github.com/vmware-tanzu/community-edition/issues/2739) - [ ] [Run AWS E2E tests for Tanzu Framework version bump PRs](https://github.com/vmware-tanzu/community-edition/issues/1519)
non_process
verify management cluster on aws is working correctly verify for management clusters on aws this issue tracks work for aws tests for management clusters verification execute run tests for management clusters this should exercise create cluster simple inspection of the cluster to make sure it s running delete cluster simple inspection to make sure the cluster is deleted is this workflow being triggered by the daily build is this workflow being triggered by modifications to makefile for a tanzu framework version bump issues found as a result of the audit we should prioritize these and determine if they are absolutely required for the workflow to function correctly and be reproducible
0
76,598
9,957,667,666
IssuesEvent
2019-07-05 17:45:18
japaf/foamgen
https://api.github.com/repos/japaf/foamgen
closed
source code documentation not build properly on RTD
documentation
There is no problem when building locally.
1.0
source code documentation not build properly on RTD - There is no problem when building locally.
non_process
source code documentation not build properly on rtd there is no problem when building locally
0
13,285
15,764,514,789
IssuesEvent
2021-03-31 13:18:55
dhh1128/ctwg
https://api.github.com/repos/dhh1128/ctwg
opened
[PROCESS] Transformation Tools
process
## Need There are two generic kinds of tools we should consider: 1. Tools that transform "ingestible documents" into the corpus internal format. 2. Tools that extract parts of the corpus in a format that can be used by (existing) tools that create presentable products, such as MkDocs, SpecUP or Docusaurus. These extraction tools select and transform data from the corpus such that they become valid input for the presentation tools. Since these tools will be used by (representatives of) existing as well as future ToIP groups, we cannot definitively predict what the restrictions should be for ingestible documents, or presentation tools, which makes finding (or creating) tools difficult. ## Proposed Solution We create a way by which the community can contribute "ingestion tools" and "extraction tools". An "ingestion tool" is 'contributed' if: - its possible configurations/settings (including defaults) are documented; - the criteria are documented that documents must satisfy in order to be ingested (= transformed into internal corpus format); - the tool exists, and can be found and operated by those that may need it. An "extraction tool" is 'contributed' if: - its possible configurations/settings (including defaults) are documented; - its extraction data model is documented (explicitly, and/or by (also) referring to the tool(s) that create presentable products (e.g. MkDocs, SpecUP, Docusaurus); - the tool exists, and can be found and operated by those that may need it. I propose to create (directories in) repos where such tools can be contributed, accepted, maintained, deprecated etc. (perhaps also with status as we use in concepts and terms). I also propose that the TT-tool be the first tool to be contributed, thereby setting an example for others to follow. It would imply that documentation is created, which might also serve as a specification for as long as the tool itself does not yet exist.
1.0
[PROCESS] Transformation Tools - ## Need There are two generic kinds of tools we should consider: 1. Tools that transform "ingestible documents" into the corpus internal format. 2. Tools that extract parts of the corpus in a format that can be used by (existing) tools that create presentable products, such as MkDocs, SpecUP or Docusaurus. These extraction tools select and transform data from the corpus such that they become valid input for the presentation tools. Since these tools will be used by (representatives of) existing as well as future ToIP groups, we cannot definitively predict what the restrictions should be for ingestible documents, or presentation tools, which makes finding (or creating) tools difficult. ## Proposed Solution We create a way by which the community can contribute "ingestion tools" and "extraction tools". An "ingestion tool" is 'contributed' if: - its possible configurations/settings (including defaults) are documented; - the criteria are documented that documents must satisfy in order to be ingested (= transformed into internal corpus format); - the tool exists, and can be found and operated by those that may need it. An "extraction tool" is 'contributed' if: - its possible configurations/settings (including defaults) are documented; - its extraction data model is documented (explicitly, and/or by (also) referring to the tool(s) that create presentable products (e.g. MkDocs, SpecUP, Docusaurus); - the tool exists, and can be found and operated by those that may need it. I propose to create (directories in) repos where such tools can be contributed, accepted, maintained, deprecated etc. (perhaps also with status as we use in concepts and terms). I also propose that the TT-tool be the first tool to be contributed, thereby setting an example for others to follow. It would imply that documentation is created, which might also serve as a specification for as long as the tool itself does not yet exist.
process
transformation tools need there are two generic kinds of tools we should consider tools that transform ingestible documents into the corpus internal format tools that extract parts of the corpus in a format that can be used by existing tools that create presentable products such as mkdocs specup or docusaurus these extraction tools select and transform data from the corpus such that they become valid input for the presentation tools since these tools will be used by representatives of existing as well as future toip groups we cannot definitively predict what the restrictions should be for ingestible documents or presentation tools which makes finding or creating tools difficult proposed solution we create a way by which the community can contribute ingestion tools and extraction tools an ingestion tool is contributed if its possible configurations settings including defaults are documented the criteria are documented that documents must satisfy in order to be ingested transformed into internal corpus format the tool exists and can be found and operated by those that may need it an extraction tool is contributed if its possible configurations settings including defaults are documented its extraction data model is documented explicitly and or by also referring to the tool s that create presentable products e g mkdocs specup docusaurus the tool exists and can be found and operated by those that may need it i propose to create directories in repos where such tools can be contributed accepted maintained deprecated etc perhaps also with status as we use in concepts and terms i also propose that the tt tool be the first tool to be contributed thereby setting an example for others to follow it would imply that documentation is created which might also serve as a specification for as long as the tool itself does not yet exist
1
429,212
30,030,299,515
IssuesEvent
2023-06-27 09:09:07
napari/docs
https://api.github.com/repos/napari/docs
opened
prevent redirects from triggering pageviews on plausible
bug documentation
## 📚 Documentation Related to napari/napari#4869, the current implementation to redirect from `/*` to `/stable/*` appears to result in triggering an event to plausible before the redirect. This complicates measuring site traffic statistics such as [Bounce Rate](https://plausible.io/docs/metrics-definitions#bounce-rate), which appears to have been artificially suppressed by the addition of redirects for versions docs. ![image](https://user-images.githubusercontent.com/1245615/181128985-3d19ca4d-43bb-491c-96d1-63b179150dc6.png) https://plausible.io/napari.org
1.0
prevent redirects from triggering pageviews on plausible - ## 📚 Documentation Related to napari/napari#4869, the current implementation to redirect from `/*` to `/stable/*` appears to result in triggering an event to plausible before the redirect. This complicates measuring site traffic statistics such as [Bounce Rate](https://plausible.io/docs/metrics-definitions#bounce-rate), which appears to have been artificially suppressed by the addition of redirects for versions docs. ![image](https://user-images.githubusercontent.com/1245615/181128985-3d19ca4d-43bb-491c-96d1-63b179150dc6.png) https://plausible.io/napari.org
non_process
prevent redirects from triggering pageviews on plausible 📚 documentation related to napari napari the current implementation to redirect from to stable appears to result in triggering an event to plausible before the redirect this complicates measuring site traffic statistics such as which appears to have been artificially suppressed by the addition of redirects for versions docs
0
14,968
18,469,572,623
IssuesEvent
2021-10-17 13:59:32
emily-writes-poems/emily-writes-poems-scripts
https://api.github.com/repos/emily-writes-poems/emily-writes-poems-scripts
closed
automate removing a poem from mongo
processing new feature
Automate these important steps for removing a poem: Gather where poem exists: - [x] from poems-list: poems that have this poem in similar poems - [x] from poems-collections-list: collections that have this poem - [x] from featured: features with this poem Confirmation and actions: - [x] This will re-run similar_poems.py to update all poems. - [x] move file into subdirectory for poems removed - [x] then run script on outer directory with all poem files - [x] requires changes to directory structure/usage - #21 - [x] This will delete record from poems-list. - [x] This will delete the poem from poem collections. - [x] Ask for confirm what to do with feature(s) -- checkbox or toggle switch: - [x] delete - [x] keep them, just set all currently_featured to false
1.0
automate removing a poem from mongo - Automate these important steps for removing a poem: Gather where poem exists: - [x] from poems-list: poems that have this poem in similar poems - [x] from poems-collections-list: collections that have this poem - [x] from featured: features with this poem Confirmation and actions: - [x] This will re-run similar_poems.py to update all poems. - [x] move file into subdirectory for poems removed - [x] then run script on outer directory with all poem files - [x] requires changes to directory structure/usage - #21 - [x] This will delete record from poems-list. - [x] This will delete the poem from poem collections. - [x] Ask for confirm what to do with feature(s) -- checkbox or toggle switch: - [x] delete - [x] keep them, just set all currently_featured to false
process
automate removing a poem from mongo automate these important steps for removing a poem gather where poem exists from poems list poems that have this poem in similar poems from poems collections list collections that have this poem from featured features with this poem confirmation and actions this will re run similar poems py to update all poems move file into subdirectory for poems removed then run script on outer directory with all poem files requires changes to directory structure usage this will delete record from poems list this will delete the poem from poem collections ask for confirm what to do with feature s checkbox or toggle switch delete keep them just set all currently featured to false
1
3,403
6,519,817,948
IssuesEvent
2017-08-28 14:17:26
openvstorage/volumedriver
https://api.github.com/repos/openvstorage/volumedriver
closed
Vpool not mount after some days
process_cantreproduce
HI all. I create a cluster on openvstorage fargo. after some days i don't work with openvstorage, vpool not mount. But process vpool still work. Thanks. Hoan
1.0
Vpool not mount after some days - HI all. I create a cluster on openvstorage fargo. after some days i don't work with openvstorage, vpool not mount. But process vpool still work. Thanks. Hoan
process
vpool not mount after some days hi all i create a cluster on openvstorage fargo after some days i don t work with openvstorage vpool not mount but process vpool still work thanks hoan
1
181,898
30,757,397,055
IssuesEvent
2023-07-29 08:22:08
DeveloperAcademy-POSTECH/MC3-Team8-Aing
https://api.github.com/repos/DeveloperAcademy-POSTECH/MC3-Team8-Aing
closed
[Design] 카메라 디자인: 툴팁 / 뒤로가기 버튼의 텍스트 / 업로드시 프로그레스 과정 텍스트로 표시
Design 클리프
## 📸 Issue <!-- 이슈에 대해 간략하게 설명해주세요 --> 카메라 부분에서 UT에서 건의사항으로 받은 디자인 부분 내용들 보강 ## 📝 To-do <!-- 진행할 작업에 대해 적어주세요 --> - [x] 사진 촬영 직후 툴팁을 보여줘서 줌, 핀치, 로테이트에 대해 인지하도록 함 (최초 1회만 표시?) - [x] 뒤로가기 버튼에 텍스트 추가 - [x] 업로드 진행시 로딩 화면에 프로그레스 과정을 텍스트로 표시 - [x] 플래시 버튼 토글 - [x] 하단 버튼 위치 조정 - [x] 전환 버튼, 라이브러리 버튼 크기 맞춤 - [x] 그 외 이슈 대응
1.0
[Design] 카메라 디자인: 툴팁 / 뒤로가기 버튼의 텍스트 / 업로드시 프로그레스 과정 텍스트로 표시 - ## 📸 Issue <!-- 이슈에 대해 간략하게 설명해주세요 --> 카메라 부분에서 UT에서 건의사항으로 받은 디자인 부분 내용들 보강 ## 📝 To-do <!-- 진행할 작업에 대해 적어주세요 --> - [x] 사진 촬영 직후 툴팁을 보여줘서 줌, 핀치, 로테이트에 대해 인지하도록 함 (최초 1회만 표시?) - [x] 뒤로가기 버튼에 텍스트 추가 - [x] 업로드 진행시 로딩 화면에 프로그레스 과정을 텍스트로 표시 - [x] 플래시 버튼 토글 - [x] 하단 버튼 위치 조정 - [x] 전환 버튼, 라이브러리 버튼 크기 맞춤 - [x] 그 외 이슈 대응
non_process
카메라 디자인 툴팁 뒤로가기 버튼의 텍스트 업로드시 프로그레스 과정 텍스트로 표시 📸 issue 카메라 부분에서 ut에서 건의사항으로 받은 디자인 부분 내용들 보강 📝 to do 사진 촬영 직후 툴팁을 보여줘서 줌 핀치 로테이트에 대해 인지하도록 함 최초 표시 뒤로가기 버튼에 텍스트 추가 업로드 진행시 로딩 화면에 프로그레스 과정을 텍스트로 표시 플래시 버튼 토글 하단 버튼 위치 조정 전환 버튼 라이브러리 버튼 크기 맞춤 그 외 이슈 대응
0
262,462
22,841,763,850
IssuesEvent
2022-07-12 22:58:32
NebulaMC-GG/Support
https://api.github.com/repos/NebulaMC-GG/Support
closed
My inventory imploded or smth ;c
support needs testing
I didn't initially delete Serene Season and some weird stuff happened to my game. Even after deleting it my inventory is still messed up in some area: I have a Shaymin form changing item that I shouldn't have, I have a Macho Brace that I shouldn't have, and I can only wear pants on my head. https://user-images.githubusercontent.com/107091311/172985468-060516e0-259a-4974-b926-1011961360cb.mp4
1.0
My inventory imploded or smth ;c - I didn't initially delete Serene Season and some weird stuff happened to my game. Even after deleting it my inventory is still messed up in some area: I have a Shaymin form changing item that I shouldn't have, I have a Macho Brace that I shouldn't have, and I can only wear pants on my head. https://user-images.githubusercontent.com/107091311/172985468-060516e0-259a-4974-b926-1011961360cb.mp4
non_process
my inventory imploded or smth c i didn t initially delete serene season and some weird stuff happened to my game even after deleting it my inventory is still messed up in some area i have a shaymin form changing item that i shouldn t have i have a macho brace that i shouldn t have and i can only wear pants on my head
0
13,782
16,540,618,223
IssuesEvent
2021-05-27 16:19:58
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
opened
Status of Bazel 5.0.0-pre.20210516.1
P1 release team-XProduct type: process
- Expected release date: Last week of May / First week of June (after 5.0.0-pre.20210510.2 from #13526) Task list: - [x] Pick release baseline: 275c54c982468b97611dc764c68daed8d7c45004 - [ ] Create release candidate: - [ ] Check downstream projects: - [ ] Push the release: - [ ] Update the [release page](https://github.com/bazelbuild/bazel/releases/)
1.0
Status of Bazel 5.0.0-pre.20210516.1 - - Expected release date: Last week of May / First week of June (after 5.0.0-pre.20210510.2 from #13526) Task list: - [x] Pick release baseline: 275c54c982468b97611dc764c68daed8d7c45004 - [ ] Create release candidate: - [ ] Check downstream projects: - [ ] Push the release: - [ ] Update the [release page](https://github.com/bazelbuild/bazel/releases/)
process
status of bazel pre expected release date last week of may first week of june after pre from task list pick release baseline create release candidate check downstream projects push the release update the
1
13,845
16,606,110,066
IssuesEvent
2021-06-02 04:09:59
AmpersandTarski/Ampersand
https://api.github.com/repos/AmpersandTarski/Ampersand
closed
Travis-ci.org will be discontinued by the end of 2020.
software process test automation
In order to maintain our automatic regression tests, we need to migrate away from Travis-ci.org. We could either go with Travis-ci.com, or leave Travis-ci and go for github actions. Here is the [documentation to migrate to Travis-ci.com](https://docs.travis-ci.com/user/migrate/open-source-repository-migration).
1.0
Travis-ci.org will be discontinued by the end of 2020. - In order to maintain our automatic regression tests, we need to migrate away from Travis-ci.org. We could either go with Travis-ci.com, or leave Travis-ci and go for github actions. Here is the [documentation to migrate to Travis-ci.com](https://docs.travis-ci.com/user/migrate/open-source-repository-migration).
process
travis ci org will be discontinued by the end of in order to maintain our automatic regression tests we need to migrate away from travis ci org we could either go with travis ci com or leave travis ci and go for github actions here is the
1
9,917
14,215,921,780
IssuesEvent
2020-11-17 08:13:36
cp-api/capella-requirements-vp
https://api.github.com/repos/cp-api/capella-requirements-vp
closed
Special characters are not correctly decoded and displayed
capella minor requirementsvp verified
Special characters (e.g. &) are not correctly decoded and displayed `ECLIPSE-555335` `POLARSYS-1382` `@tto` `2017-01-09`
1.0
Special characters are not correctly decoded and displayed - Special characters (e.g. &) are not correctly decoded and displayed `ECLIPSE-555335` `POLARSYS-1382` `@tto` `2017-01-09`
non_process
special characters are not correctly decoded and displayed special characters e g are not correctly decoded and displayed eclipse polarsys tto
0
16,577
21,607,319,493
IssuesEvent
2022-05-04 05:57:49
arcus-azure/arcus.messaging
https://api.github.com/repos/arcus-azure/arcus.messaging
closed
Add support for open message handler interface extraction in Azure Functions .NET 6
enhancement dependencies area:message-processing
**Is your feature request related to a problem? Please describe.** We use open interfaces for message handlers, which are now supported in Azure Functions .NET Core, but not yet in .NET 6. This is probably because of internal types that are renamed. **Describe the solution you'd like** Make sure that we support .NET 6 types when extracting open message handler interfaces. This can be done by upgrading the Azure Functions test project.
1.0
Add support for open message handler interface extraction in Azure Functions .NET 6 - **Is your feature request related to a problem? Please describe.** We use open interfaces for message handlers, which are now supported in Azure Functions .NET Core, but not yet in .NET 6. This is probably because of internal types that are renamed. **Describe the solution you'd like** Make sure that we support .NET 6 types when extracting open message handler interfaces. This can be done by upgrading the Azure Functions test project.
process
add support for open message handler interface extraction in azure functions net is your feature request related to a problem please describe we use open interfaces for message handlers which are now supported in azure functions net core but not yet in net this is probably because of internal types that are renamed describe the solution you d like make sure that we support net types when extracting open message handler interfaces this can be done by upgrading the azure functions test project
1
190,140
22,047,245,826
IssuesEvent
2022-05-30 04:09:45
nanopathi/linux-4.19.72_CVE-2021-32399
https://api.github.com/repos/nanopathi/linux-4.19.72_CVE-2021-32399
closed
CVE-2020-25705 (High) detected in linuxlinux-4.19.236 - autoclosed
security vulnerability
## CVE-2020-25705 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.236</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nanopathi/linux-4.19.72_CVE-2021-32399/commit/03cb3c6f0e0b62b5cbcd747df63781fbb2a6ef66">03cb3c6f0e0b62b5cbcd747df63781fbb2a6ef66</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/ipv4/icmp.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw in ICMP packets in the Linux kernel may allow an attacker to quickly scan open UDP ports. This flaw allows an off-path remote attacker to effectively bypass source port UDP randomization. Software that relies on UDP source port randomization are indirectly affected as well on the Linux Based Products (RUGGEDCOM RM1224: All versions between v5.0 and v6.4, SCALANCE M-800: All versions between v5.0 and v6.4, SCALANCE S615: All versions between v5.0 and v6.4, SCALANCE SC-600: All versions prior to v2.1.3, SCALANCE W1750D: v8.3.0.1, v8.6.0, and v8.7.0, SIMATIC Cloud Connect 7: All versions, SIMATIC MV500 Family: All versions, SIMATIC NET CP 1243-1 (incl. SIPLUS variants): Versions 3.1.39 and later, SIMATIC NET CP 1243-7 LTE EU: Version <p>Publish Date: 2020-11-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-25705>CVE-2020-25705</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-25705">https://www.linuxkernelcves.com/cves/CVE-2020-25705</a></p> <p>Release Date: 2020-11-17</p> <p>Fix Resolution: v4.4.241,v4.9.241,v4.14.203,v4.19.153,v5.4.73,v5.8.17,v5.9.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-25705 (High) detected in linuxlinux-4.19.236 - autoclosed - ## CVE-2020-25705 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.236</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/nanopathi/linux-4.19.72_CVE-2021-32399/commit/03cb3c6f0e0b62b5cbcd747df63781fbb2a6ef66">03cb3c6f0e0b62b5cbcd747df63781fbb2a6ef66</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/ipv4/icmp.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A flaw in ICMP packets in the Linux kernel may allow an attacker to quickly scan open UDP ports. This flaw allows an off-path remote attacker to effectively bypass source port UDP randomization. Software that relies on UDP source port randomization are indirectly affected as well on the Linux Based Products (RUGGEDCOM RM1224: All versions between v5.0 and v6.4, SCALANCE M-800: All versions between v5.0 and v6.4, SCALANCE S615: All versions between v5.0 and v6.4, SCALANCE SC-600: All versions prior to v2.1.3, SCALANCE W1750D: v8.3.0.1, v8.6.0, and v8.7.0, SIMATIC Cloud Connect 7: All versions, SIMATIC MV500 Family: All versions, SIMATIC NET CP 1243-1 (incl. SIPLUS variants): Versions 3.1.39 and later, SIMATIC NET CP 1243-7 LTE EU: Version <p>Publish Date: 2020-11-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-25705>CVE-2020-25705</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-25705">https://www.linuxkernelcves.com/cves/CVE-2020-25705</a></p> <p>Release Date: 2020-11-17</p> <p>Fix Resolution: v4.4.241,v4.9.241,v4.14.203,v4.19.153,v5.4.73,v5.8.17,v5.9.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in linuxlinux autoclosed cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files net icmp c vulnerability details a flaw in icmp packets in the linux kernel may allow an attacker to quickly scan open udp ports this flaw allows an off path remote attacker to effectively bypass source port udp randomization software that relies on udp source port randomization are indirectly affected as well on the linux based products ruggedcom all versions between and scalance m all versions between and scalance all versions between and scalance sc all versions prior to scalance and simatic cloud connect all versions simatic family all versions simatic net cp incl siplus variants versions and later simatic net cp lte eu version publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
150,871
23,725,911,133
IssuesEvent
2022-08-30 19:35:13
dotnet/roslyn
https://api.github.com/repos/dotnet/roslyn
reopened
Missing CS8166 when accessing [UnscopedRef] property to assing to an unscoped local
Area-Compilers Resolution-By Design
**Version Used**: 4.4.0-2.22419.1 (ae566380) **Steps to Reproduce**: 1. Paste the following code: ```csharp ref struct A { [UnscopedRef] public Span<byte> Span => default; } class Foo { void Bar(ref A a) { Span<byte> span = a.Span; } } ``` **Expected Behavior**: The code should produce a CS8166 error (same as #63524, but correctly here), because `span` is not `scoped`. **Actual Behavior**: Code compiles fine. ![](https://user-images.githubusercontent.com/10199417/186033466-46973d7b-cf8e-4da6-9083-40802aa121a4.png) cc. @cston
1.0
Missing CS8166 when accessing [UnscopedRef] property to assing to an unscoped local - **Version Used**: 4.4.0-2.22419.1 (ae566380) **Steps to Reproduce**: 1. Paste the following code: ```csharp ref struct A { [UnscopedRef] public Span<byte> Span => default; } class Foo { void Bar(ref A a) { Span<byte> span = a.Span; } } ``` **Expected Behavior**: The code should produce a CS8166 error (same as #63524, but correctly here), because `span` is not `scoped`. **Actual Behavior**: Code compiles fine. ![](https://user-images.githubusercontent.com/10199417/186033466-46973d7b-cf8e-4da6-9083-40802aa121a4.png) cc. @cston
non_process
missing when accessing property to assing to an unscoped local version used steps to reproduce paste the following code csharp ref struct a public span span default class foo void bar ref a a span span a span expected behavior the code should produce a error same as but correctly here because span is not scoped actual behavior code compiles fine cc cston
0
54,415
7,887,440,873
IssuesEvent
2018-06-27 18:29:09
GoogleCloudPlatform/google-cloud-java
https://api.github.com/repos/GoogleCloudPlatform/google-cloud-java
closed
Document in javadoc that certain BigQuery methods do not interact with streaming buffer
api: bigquery documentation priority: p2
So issues like https://github.com/GoogleCloudPlatform/google-cloud-java/issues/3344 can be immediately be avoided by users.
1.0
Document in javadoc that certain BigQuery methods do not interact with streaming buffer - So issues like https://github.com/GoogleCloudPlatform/google-cloud-java/issues/3344 can be immediately be avoided by users.
non_process
document in javadoc that certain bigquery methods do not interact with streaming buffer so issues like can be immediately be avoided by users
0
3,727
6,733,142,279
IssuesEvent
2017-10-18 13:58:37
york-region-tpss/stp
https://api.github.com/repos/york-region-tpss/stp
closed
checkbox and Lock selected functionality
enhancement process workflow report
Add apex_item checkbox and lock selected button on contract preparation dashboard so that user can lock items in bulk. The button locates at the right corner of report search bar.
1.0
checkbox and Lock selected functionality - Add apex_item checkbox and lock selected button on contract preparation dashboard so that user can lock items in bulk. The button locates at the right corner of report search bar.
process
checkbox and lock selected functionality add apex item checkbox and lock selected button on contract preparation dashboard so that user can lock items in bulk the button locates at the right corner of report search bar
1
66,033
3,249,684,378
IssuesEvent
2015-10-18 11:27:24
mattmezza/socialize
https://api.github.com/repos/mattmezza/socialize
opened
PUT /posts/:id_post/comments/:id_comment/replies/:id_reply/like
API low-priority
This api likes a reply (which is a comment to a comment actually).
1.0
PUT /posts/:id_post/comments/:id_comment/replies/:id_reply/like - This api likes a reply (which is a comment to a comment actually).
non_process
put posts id post comments id comment replies id reply like this api likes a reply which is a comment to a comment actually
0
120,458
4,789,502,377
IssuesEvent
2016-10-31 01:57:22
OctopusDeploy/Issues
https://api.github.com/repos/OctopusDeploy/Issues
closed
Steps are mistakenly excluded from auto-deploy when a Step is scoped to multiple roles
bug feature/elastictransientenvironments priority release/3.4
When an automatic deployment is triggered, the resulting steps that execute should be based on the roles of the machines that are included in the automatic deployment with the primary goal of making the new machine just like all the other machines with the same roles. If a step is configured to run for multiple roles, the step will be mistakenly excluded from an auto-deploy when the target machine doesn't have ALL of the roles configured on the step. It should actually be "run this step on the target machine if the target machine has ANY of these roles".
1.0
Steps are mistakenly excluded from auto-deploy when a Step is scoped to multiple roles - When an automatic deployment is triggered, the resulting steps that execute should be based on the roles of the machines that are included in the automatic deployment with the primary goal of making the new machine just like all the other machines with the same roles. If a step is configured to run for multiple roles, the step will be mistakenly excluded from an auto-deploy when the target machine doesn't have ALL of the roles configured on the step. It should actually be "run this step on the target machine if the target machine has ANY of these roles".
non_process
steps are mistakenly excluded from auto deploy when a step is scoped to multiple roles when an automatic deployment is triggered the resulting steps that execute should be based on the roles of the machines that are included in the automatic deployment with the primary goal of making the new machine just like all the other machines with the same roles if a step is configured to run for multiple roles the step will be mistakenly excluded from an auto deploy when the target machine doesn t have all of the roles configured on the step it should actually be run this step on the target machine if the target machine has any of these roles
0
4,654
7,496,143,734
IssuesEvent
2018-04-08 06:06:11
kookmin-sw/2018-cap1-2
https://api.github.com/repos/kookmin-sw/2018-cap1-2
closed
파이썬 영상처리 -2- 이미지 잡음 처리 및 명확하게 전처리
ImageProcessing
현재까지의 코드로는 글자 이미지와 종이가 명확하게 구별이 잘 되지 않음. 또한 히스토그램 분포가 들어오는 이미지의 종류에 따라 다양하여 일괄적으로 임계값을 적용할 수 없음. 모든 이미지에 대하여 일정한 임계값으로 처리할 수 있어야 자동화가 성립됨.
1.0
파이썬 영상처리 -2- 이미지 잡음 처리 및 명확하게 전처리 - 현재까지의 코드로는 글자 이미지와 종이가 명확하게 구별이 잘 되지 않음. 또한 히스토그램 분포가 들어오는 이미지의 종류에 따라 다양하여 일괄적으로 임계값을 적용할 수 없음. 모든 이미지에 대하여 일정한 임계값으로 처리할 수 있어야 자동화가 성립됨.
process
파이썬 영상처리 이미지 잡음 처리 및 명확하게 전처리 현재까지의 코드로는 글자 이미지와 종이가 명확하게 구별이 잘 되지 않음 또한 히스토그램 분포가 들어오는 이미지의 종류에 따라 다양하여 일괄적으로 임계값을 적용할 수 없음 모든 이미지에 대하여 일정한 임계값으로 처리할 수 있어야 자동화가 성립됨
1
508,463
14,700,810,602
IssuesEvent
2021-01-04 10:48:16
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.eporner.com - video or audio doesn't play
browser-firefox-mobile engine-gecko ml-needsdiagnosis-false priority-normal
<!-- @browser: Firefox Mobile 81.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:81.0) Gecko/81.0 Firefox/81.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/64841 --> **URL**: https://www.eporner.com/video-ILjwuLoTDjG/a-random-wanker-wants-to-join-in/ **Browser / Version**: Firefox Mobile 81.0 **Operating System**: Android **Tested Another Browser**: No **Problem type**: Video or audio doesn't play **Description**: The video or audio does not play **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200804091327</li><li>channel: nightly</li><li>hasTouchScreen: true</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.eporner.com - video or audio doesn't play - <!-- @browser: Firefox Mobile 81.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:81.0) Gecko/81.0 Firefox/81.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/64841 --> **URL**: https://www.eporner.com/video-ILjwuLoTDjG/a-random-wanker-wants-to-join-in/ **Browser / Version**: Firefox Mobile 81.0 **Operating System**: Android **Tested Another Browser**: No **Problem type**: Video or audio doesn't play **Description**: The video or audio does not play **Steps to Reproduce**: <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200804091327</li><li>channel: nightly</li><li>hasTouchScreen: true</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
video or audio doesn t play url browser version firefox mobile operating system android tested another browser no problem type video or audio doesn t play description the video or audio does not play steps to reproduce browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true from with ❤️
0
127,837
27,138,212,305
IssuesEvent
2023-02-16 14:39:36
TheOpenSpaceProgram/new-ospgl
https://api.github.com/repos/TheOpenSpaceProgram/new-ospgl
closed
Allow specifying sizes for child_1 in canvas
enhancement good first issue code-improvement
Right now we only allow size specification in pixels for child_0 because of how the code was structured originally, it should now be possible to specify size for child_1.
1.0
Allow specifying sizes for child_1 in canvas - Right now we only allow size specification in pixels for child_0 because of how the code was structured originally, it should now be possible to specify size for child_1.
non_process
allow specifying sizes for child in canvas right now we only allow size specification in pixels for child because of how the code was structured originally it should now be possible to specify size for child
0
15,524
19,703,269,010
IssuesEvent
2022-01-12 18:52:27
googleapis/java-gsuite-addons
https://api.github.com/repos/googleapis/java-gsuite-addons
opened
Your .repo-metadata.json file has a problem 🤒
type: process repo-metadata: lint
You have a problem with your .repo-metadata.json file: Result of scan 📈: * release_level must be equal to one of the allowed values in .repo-metadata.json * api_shortname 'gsuite-addons' invalid in .repo-metadata.json ☝️ Once you correct these problems, you can close this issue. Reach out to **go/github-automation** if you have any questions.
1.0
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file: Result of scan 📈: * release_level must be equal to one of the allowed values in .repo-metadata.json * api_shortname 'gsuite-addons' invalid in .repo-metadata.json ☝️ Once you correct these problems, you can close this issue. Reach out to **go/github-automation** if you have any questions.
process
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 release level must be equal to one of the allowed values in repo metadata json api shortname gsuite addons invalid in repo metadata json ☝️ once you correct these problems you can close this issue reach out to go github automation if you have any questions
1
213,021
16,507,178,451
IssuesEvent
2021-05-25 20:53:41
ABCD-STUDY/nda-abcd-collection-3165
https://api.github.com/repos/ABCD-STUDY/nda-abcd-collection-3165
closed
Connectivity Matrix download
documentation high priority question
Hi, I downloaded the Gordon atlas connecitvity matrices and noticed that some subjects does not have all the three types (5min, 10min and below threshold). Is it resonable in case the subject has less number of valid frames, e.g. do have a 5min matrix but not a 10min matrix. Can you please explain why would a subject have a 10min matrix but not a 5min matrix?
1.0
Connectivity Matrix download - Hi, I downloaded the Gordon atlas connecitvity matrices and noticed that some subjects does not have all the three types (5min, 10min and below threshold). Is it resonable in case the subject has less number of valid frames, e.g. do have a 5min matrix but not a 10min matrix. Can you please explain why would a subject have a 10min matrix but not a 5min matrix?
non_process
connectivity matrix download hi i downloaded the gordon atlas connecitvity matrices and noticed that some subjects does not have all the three types and below threshold is it resonable in case the subject has less number of valid frames e g do have a matrix but not a matrix can you please explain why would a subject have a matrix but not a matrix
0
313,954
26,965,060,068
IssuesEvent
2023-02-08 21:33:17
rancher/dashboard
https://api.github.com/repos/rancher/dashboard
closed
[EKS] Support custom instance roles for node groups
[zube]: To Test internal priority/0 QA/XS kind/enhancement team/area2 JIRA
In order to support custom node instance roles for EKS clusters, a new dropdown should be added to the "Add nodegroup" section of the UI. This dropdown should: 1. Get the available node instance roles from the AWS API to populate the dropdown. 2. Filter out the node instance roles that are prefixed with `rancher-managed-` 3. The field name should be `nodeInstanceRole` 4. The value passed to the Rancher API should be a `string`. If no role is selected for a nodegroup, then the empty string should be passed. SURE-4536
1.0
[EKS] Support custom instance roles for node groups - In order to support custom node instance roles for EKS clusters, a new dropdown should be added to the "Add nodegroup" section of the UI. This dropdown should: 1. Get the available node instance roles from the AWS API to populate the dropdown. 2. Filter out the node instance roles that are prefixed with `rancher-managed-` 3. The field name should be `nodeInstanceRole` 4. The value passed to the Rancher API should be a `string`. If no role is selected for a nodegroup, then the empty string should be passed. SURE-4536
non_process
support custom instance roles for node groups in order to support custom node instance roles for eks clusters a new dropdown should be added to the add nodegroup section of the ui this dropdown should get the available node instance roles from the aws api to populate the dropdown filter out the node instance roles that are prefixed with rancher managed the field name should be nodeinstancerole the value passed to the rancher api should be a string if no role is selected for a nodegroup then the empty string should be passed sure
0
4,323
7,227,916,794
IssuesEvent
2018-02-11 02:27:52
hardvolk/foodie-journal
https://api.github.com/repos/hardvolk/foodie-journal
closed
Integrar pantallas con Router
In process
Actualmente las pantallas están en componentes individuales. Utilizar el módulo de Router y definir las rutas. Posteriormente integrar los componentes por medio de sus rutas. Si no se ha iniciado, me puedo encargar de ello.
1.0
Integrar pantallas con Router - Actualmente las pantallas están en componentes individuales. Utilizar el módulo de Router y definir las rutas. Posteriormente integrar los componentes por medio de sus rutas. Si no se ha iniciado, me puedo encargar de ello.
process
integrar pantallas con router actualmente las pantallas están en componentes individuales utilizar el módulo de router y definir las rutas posteriormente integrar los componentes por medio de sus rutas si no se ha iniciado me puedo encargar de ello
1
12,407
14,916,286,048
IssuesEvent
2021-01-22 17:58:19
open-telemetry/opentelemetry-collector
https://api.github.com/repos/open-telemetry/opentelemetry-collector
closed
Bug report for failed CircleCI build (job: test)
area:exporter area:processor bug priority:p3 release:allowed-for-ga spec:logs spec:metrics spec:trace
Auto-generated report for `test` job build. Link to failed build: https://circleci.com/gh/open-telemetry/opentelemetry-collector/49855 #### Test Failures - TestQueuedRetry_StopWhileWaiting **Note**: Information about any subsequent build failures that happen while this issue is open, will be added as comments with more information to this issue.
1.0
Bug report for failed CircleCI build (job: test) - Auto-generated report for `test` job build. Link to failed build: https://circleci.com/gh/open-telemetry/opentelemetry-collector/49855 #### Test Failures - TestQueuedRetry_StopWhileWaiting **Note**: Information about any subsequent build failures that happen while this issue is open, will be added as comments with more information to this issue.
process
bug report for failed circleci build job test auto generated report for test job build link to failed build test failures testqueuedretry stopwhilewaiting note information about any subsequent build failures that happen while this issue is open will be added as comments with more information to this issue
1
18,517
24,551,735,061
IssuesEvent
2022-10-12 13:07:11
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[iOS] Only first consent is getting displayed to the mobile participants
Bug Blocker P0 iOS Process: Fixed Process: Tested dev
Steps: 1. In SB, click on create study 2. In the consent section, add multiple consents and launch the study 3. Go to the mobile app 4. Sign up or sign in to the mobile app 5. Click on the created study 6. while enrolling, observe the consent AR: Only first consent is getting displayed to the mobile participants ER: All the created consent should get displayed to the mobile participants **Configurations in SB** ![image](https://user-images.githubusercontent.com/71445210/186197508-bbb9d28a-ad7c-44a2-bf08-75b62de65d28.png) https://user-images.githubusercontent.com/71445210/186198635-b1b864b4-4e69-4727-b3c5-22ec875f8b5b.MOV
2.0
[iOS] Only first consent is getting displayed to the mobile participants - Steps: 1. In SB, click on create study 2. In the consent section, add multiple consents and launch the study 3. Go to the mobile app 4. Sign up or sign in to the mobile app 5. Click on the created study 6. while enrolling, observe the consent AR: Only first consent is getting displayed to the mobile participants ER: All the created consent should get displayed to the mobile participants **Configurations in SB** ![image](https://user-images.githubusercontent.com/71445210/186197508-bbb9d28a-ad7c-44a2-bf08-75b62de65d28.png) https://user-images.githubusercontent.com/71445210/186198635-b1b864b4-4e69-4727-b3c5-22ec875f8b5b.MOV
process
only first consent is getting displayed to the mobile participants steps in sb click on create study in the consent section add multiple consents and launch the study go to the mobile app sign up or sign in to the mobile app click on the created study while enrolling observe the consent ar only first consent is getting displayed to the mobile participants er all the created consent should get displayed to the mobile participants configurations in sb
1
22,301
30,857,267,221
IssuesEvent
2023-08-02 21:50:40
NCAR/ucomp-pipeline
https://api.github.com/repos/NCAR/ucomp-pipeline
opened
Bad mean/median files on 2022-02-02
process
@bberkeyU: > For Feb 2-2-22 All the 3 point L1 data looks good but the one 5 point image (which looks like it has the wrong wavelength off set) is killing the median results
1.0
Bad mean/median files on 2022-02-02 - @bberkeyU: > For Feb 2-2-22 All the 3 point L1 data looks good but the one 5 point image (which looks like it has the wrong wavelength off set) is killing the median results
process
bad mean median files on bberkeyu for feb all the point data looks good but the one point image which looks like it has the wrong wavelength off set is killing the median results
1
129,380
17,775,750,738
IssuesEvent
2021-08-30 18:59:29
CMSgov/eAPD
https://api.github.com/repos/CMSgov/eAPD
closed
[Design Issue] Design a pattern for the export view for when non-required fields are not filled out
design Small
Some of the fields in the eAPD will not be required to be answered before submit. We have language that we use, noted in the content guide in the wiki, to note when a field isn't filled out so that it's clear to the reviewer that it wasn't a mistake or something broken in the system, but we only did this for top fields before release. Determine if we need any more language patterns or visual patterns to show when fields are not answered, when we include ALL fields in the list. ### This task is done when… - [ ] any acceptance criteria (not process oriented, requirements of feature) - [x] designs are created, taking into consideration multiple FFY, multiple entries, long entries, weird entries, and/or the export view if applicable, validation considerations - [ ] 508 considerations and/or recommendations are reviewed - [x] designs are shared - [ ] designs are approved by product - [ ] if there is a long discussion, a summary of changes for dev will be added to the last comment before transfer - [ ] a list of changes to be made if this is improving an existing feature --- Add additional labels (design, dev, compliance, BUG, etc) and size before submitting. If the issue is needed to complete prioritized work for the CURRENT SPRINT, add it to the "This Sprint" pipeline. Otherwise, all other issues will be automatically added to the _unprioritized pipeline_ for prioritization in backlog refinement or sprint planning with Product.
1.0
[Design Issue] Design a pattern for the export view for when non-required fields are not filled out - Some of the fields in the eAPD will not be required to be answered before submit. We have language that we use, noted in the content guide in the wiki, to note when a field isn't filled out so that it's clear to the reviewer that it wasn't a mistake or something broken in the system, but we only did this for top fields before release. Determine if we need any more language patterns or visual patterns to show when fields are not answered, when we include ALL fields in the list. ### This task is done when… - [ ] any acceptance criteria (not process oriented, requirements of feature) - [x] designs are created, taking into consideration multiple FFY, multiple entries, long entries, weird entries, and/or the export view if applicable, validation considerations - [ ] 508 considerations and/or recommendations are reviewed - [x] designs are shared - [ ] designs are approved by product - [ ] if there is a long discussion, a summary of changes for dev will be added to the last comment before transfer - [ ] a list of changes to be made if this is improving an existing feature --- Add additional labels (design, dev, compliance, BUG, etc) and size before submitting. If the issue is needed to complete prioritized work for the CURRENT SPRINT, add it to the "This Sprint" pipeline. Otherwise, all other issues will be automatically added to the _unprioritized pipeline_ for prioritization in backlog refinement or sprint planning with Product.
non_process
design a pattern for the export view for when non required fields are not filled out some of the fields in the eapd will not be required to be answered before submit we have language that we use noted in the content guide in the wiki to note when a field isn t filled out so that it s clear to the reviewer that it wasn t a mistake or something broken in the system but we only did this for top fields before release determine if we need any more language patterns or visual patterns to show when fields are not answered when we include all fields in the list this task is done when… any acceptance criteria not process oriented requirements of feature designs are created taking into consideration multiple ffy multiple entries long entries weird entries and or the export view if applicable validation considerations considerations and or recommendations are reviewed designs are shared designs are approved by product if there is a long discussion a summary of changes for dev will be added to the last comment before transfer a list of changes to be made if this is improving an existing feature add additional labels design dev compliance bug etc and size before submitting if the issue is needed to complete prioritized work for the current sprint add it to the this sprint pipeline otherwise all other issues will be automatically added to the unprioritized pipeline for prioritization in backlog refinement or sprint planning with product
0
161,284
13,822,258,163
IssuesEvent
2020-10-13 04:37:26
chriswalz/bit
https://api.github.com/repos/chriswalz/bit
closed
What does 'bit release' do?
documentation good first issue
'Generate a production release' doesn't say very much. Here is the output from running `bit release` offline: ``` $ bit release Error: accepts 1 arg(s), received 0 $ bit release a exit status 1 ## readme...personal/readme ? Please provide a description of your changes [Enter 2 empty lines to finish] # I hit ctrl+c, because I don't know whether it's referring to the most recent commit or all commits since the latest release On branch readme Your branch is up to date with 'personal/readme'. Untracked files: (use "git add <file>..." to include in what will be committed) ... nothing added to commit but untracked files present (use "git add" to track) ssh: Could not resolve hostname github.com: Temporary failure in name resolution fatal: Could not read from remote repository. Please make sure you have the correct access rights and the repository exists. ssh: Could not resolve hostname github.com: Temporary failure in name resolution fatal: Could not read from remote repository. Please make sure you have the correct access rights and the repository exists. ``` If `bit` is going to perform arbitrary git operations, I'd like to at least know what they are.
1.0
What does 'bit release' do? - 'Generate a production release' doesn't say very much. Here is the output from running `bit release` offline: ``` $ bit release Error: accepts 1 arg(s), received 0 $ bit release a exit status 1 ## readme...personal/readme ? Please provide a description of your changes [Enter 2 empty lines to finish] # I hit ctrl+c, because I don't know whether it's referring to the most recent commit or all commits since the latest release On branch readme Your branch is up to date with 'personal/readme'. Untracked files: (use "git add <file>..." to include in what will be committed) ... nothing added to commit but untracked files present (use "git add" to track) ssh: Could not resolve hostname github.com: Temporary failure in name resolution fatal: Could not read from remote repository. Please make sure you have the correct access rights and the repository exists. ssh: Could not resolve hostname github.com: Temporary failure in name resolution fatal: Could not read from remote repository. Please make sure you have the correct access rights and the repository exists. ``` If `bit` is going to perform arbitrary git operations, I'd like to at least know what they are.
non_process
what does bit release do generate a production release doesn t say very much here is the output from running bit release offline bit release error accepts arg s received bit release a exit status readme personal readme please provide a description of your changes i hit ctrl c because i don t know whether it s referring to the most recent commit or all commits since the latest release on branch readme your branch is up to date with personal readme untracked files use git add to include in what will be committed nothing added to commit but untracked files present use git add to track ssh could not resolve hostname github com temporary failure in name resolution fatal could not read from remote repository please make sure you have the correct access rights and the repository exists ssh could not resolve hostname github com temporary failure in name resolution fatal could not read from remote repository please make sure you have the correct access rights and the repository exists if bit is going to perform arbitrary git operations i d like to at least know what they are
0
367,670
10,860,458,814
IssuesEvent
2019-11-14 09:07:05
googleapis/java-gcloud-maven-plugin
https://api.github.com/repos/googleapis/java-gcloud-maven-plugin
opened
Synthesis failed for java-gcloud-maven-plugin
autosynth failure priority: p1 type: bug
Hello! Autosynth couldn't regenerate java-gcloud-maven-plugin. :broken_heart: Here's the output from running `synth.py`: ``` Cloning into 'working_repo'... Switched to branch 'autosynth' Running synthtool ['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', 'synth.py', '--'] synthtool > Executing /tmpfs/src/git/autosynth/working_repo/synth.py. .github/ISSUE_TEMPLATE/bug_report.md .github/ISSUE_TEMPLATE/feature_request.md .github/ISSUE_TEMPLATE/support_request.md .github/PULL_REQUEST_TEMPLATE.md .github/release-please.yml .kokoro/build.bat .kokoro/build.sh .kokoro/coerce_logs.sh .kokoro/common.cfg .kokoro/continuous/common.cfg .kokoro/continuous/dependencies.cfg .kokoro/continuous/integration.cfg .kokoro/continuous/java11.cfg .kokoro/continuous/java7.cfg .kokoro/continuous/java8-osx.cfg .kokoro/continuous/java8-win.cfg .kokoro/continuous/java8.cfg .kokoro/continuous/lint.cfg .kokoro/continuous/propose_release.cfg .kokoro/continuous/propose_release.sh .kokoro/dependencies.sh .kokoro/linkage-monitor.sh .kokoro/nightly/common.cfg .kokoro/nightly/dependencies.cfg .kokoro/nightly/integration.cfg .kokoro/nightly/java11.cfg .kokoro/nightly/java7.cfg .kokoro/nightly/java8-osx.cfg .kokoro/nightly/java8-win.cfg .kokoro/nightly/java8.cfg .kokoro/nightly/lint.cfg .kokoro/presubmit/clirr.cfg .kokoro/presubmit/common.cfg .kokoro/presubmit/dependencies.cfg .kokoro/presubmit/integration.cfg .kokoro/presubmit/java11.cfg .kokoro/presubmit/java7.cfg .kokoro/presubmit/java8-osx.cfg .kokoro/presubmit/java8-win.cfg .kokoro/presubmit/java8.cfg .kokoro/presubmit/linkage-monitor.cfg .kokoro/presubmit/lint.cfg .kokoro/release/bump_snapshot.cfg .kokoro/release/bump_snapshot.sh .kokoro/release/common.cfg .kokoro/release/common.sh .kokoro/release/drop.cfg .kokoro/release/drop.sh .kokoro/release/promote.cfg .kokoro/release/promote.sh .kokoro/release/publish_javadoc.cfg .kokoro/release/publish_javadoc.sh Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/__main__.py", line 87, in <module> main() File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 764, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 717, in main rv = self.invoke(ctx) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 956, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 555, in invoke return callback(*args, **kwargs) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/__main__.py", line 79, in main spec.loader.exec_module(synth_module) # type: ignore File "<frozen importlib._bootstrap_external>", line 678, in exec_module File "<frozen importlib._bootstrap>", line 205, in _call_with_frames_removed File "/tmpfs/src/git/autosynth/working_repo/synth.py", line 21, in <module> templates = common_templates.java_library() File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/gcp/common.py", line 63, in java_library return self._generic_library("java_library", **kwargs) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/gcp/common.py", line 49, in _generic_library result = t.render(**kwargs) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/sources/templates.py", line 83, in render _render_to_path(self.env, template_name, self.dir, kwargs) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/sources/templates.py", line 53, in _render_to_path output.dump(fh) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/jinja2/environment.py", line 1227, in dump fp.writelines(iterable) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/jinja2/environment.py", line 1271, in __next__ return self._next() File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/jinja2/environment.py", line 1045, in generate yield self.environment.handle_exception(exc_info, True) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/jinja2/environment.py", line 780, in handle_exception reraise(exc_type, exc_value, tb) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/jinja2/_compat.py", line 37, in reraise raise value.with_traceback(tb) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/gcp/templates/java_library/.kokoro/release/publish_javadoc.sh", line 36, in top-level template code NAME={{ metadata['repo']['distribution_name'].split(':')|last }} File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/jinja2/environment.py", line 430, in getattr return getattr(obj, attribute) jinja2.exceptions.UndefinedError: 'dict object' has no attribute 'distribution_name' synthtool > Cleaned up 2 temporary directories. synthtool > Wrote metadata to synth.metadata. Synthesis failed ``` Google internal developers can see the full log [here](https://sponge/f2df4be1-2bc3-402c-9ee3-d298858423e6).
1.0
Synthesis failed for java-gcloud-maven-plugin - Hello! Autosynth couldn't regenerate java-gcloud-maven-plugin. :broken_heart: Here's the output from running `synth.py`: ``` Cloning into 'working_repo'... Switched to branch 'autosynth' Running synthtool ['/tmpfs/src/git/autosynth/env/bin/python3', '-m', 'synthtool', 'synth.py', '--'] synthtool > Executing /tmpfs/src/git/autosynth/working_repo/synth.py. .github/ISSUE_TEMPLATE/bug_report.md .github/ISSUE_TEMPLATE/feature_request.md .github/ISSUE_TEMPLATE/support_request.md .github/PULL_REQUEST_TEMPLATE.md .github/release-please.yml .kokoro/build.bat .kokoro/build.sh .kokoro/coerce_logs.sh .kokoro/common.cfg .kokoro/continuous/common.cfg .kokoro/continuous/dependencies.cfg .kokoro/continuous/integration.cfg .kokoro/continuous/java11.cfg .kokoro/continuous/java7.cfg .kokoro/continuous/java8-osx.cfg .kokoro/continuous/java8-win.cfg .kokoro/continuous/java8.cfg .kokoro/continuous/lint.cfg .kokoro/continuous/propose_release.cfg .kokoro/continuous/propose_release.sh .kokoro/dependencies.sh .kokoro/linkage-monitor.sh .kokoro/nightly/common.cfg .kokoro/nightly/dependencies.cfg .kokoro/nightly/integration.cfg .kokoro/nightly/java11.cfg .kokoro/nightly/java7.cfg .kokoro/nightly/java8-osx.cfg .kokoro/nightly/java8-win.cfg .kokoro/nightly/java8.cfg .kokoro/nightly/lint.cfg .kokoro/presubmit/clirr.cfg .kokoro/presubmit/common.cfg .kokoro/presubmit/dependencies.cfg .kokoro/presubmit/integration.cfg .kokoro/presubmit/java11.cfg .kokoro/presubmit/java7.cfg .kokoro/presubmit/java8-osx.cfg .kokoro/presubmit/java8-win.cfg .kokoro/presubmit/java8.cfg .kokoro/presubmit/linkage-monitor.cfg .kokoro/presubmit/lint.cfg .kokoro/release/bump_snapshot.cfg .kokoro/release/bump_snapshot.sh .kokoro/release/common.cfg .kokoro/release/common.sh .kokoro/release/drop.cfg .kokoro/release/drop.sh .kokoro/release/promote.cfg .kokoro/release/promote.sh .kokoro/release/publish_javadoc.cfg .kokoro/release/publish_javadoc.sh Traceback (most recent call last): File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 193, in _run_module_as_main "__main__", mod_spec) File "/home/kbuilder/.pyenv/versions/3.6.1/lib/python3.6/runpy.py", line 85, in _run_code exec(code, run_globals) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/__main__.py", line 87, in <module> main() File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 764, in __call__ return self.main(*args, **kwargs) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 717, in main rv = self.invoke(ctx) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 956, in invoke return ctx.invoke(self.callback, **ctx.params) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/click/core.py", line 555, in invoke return callback(*args, **kwargs) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/__main__.py", line 79, in main spec.loader.exec_module(synth_module) # type: ignore File "<frozen importlib._bootstrap_external>", line 678, in exec_module File "<frozen importlib._bootstrap>", line 205, in _call_with_frames_removed File "/tmpfs/src/git/autosynth/working_repo/synth.py", line 21, in <module> templates = common_templates.java_library() File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/gcp/common.py", line 63, in java_library return self._generic_library("java_library", **kwargs) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/gcp/common.py", line 49, in _generic_library result = t.render(**kwargs) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/sources/templates.py", line 83, in render _render_to_path(self.env, template_name, self.dir, kwargs) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/sources/templates.py", line 53, in _render_to_path output.dump(fh) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/jinja2/environment.py", line 1227, in dump fp.writelines(iterable) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/jinja2/environment.py", line 1271, in __next__ return self._next() File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/jinja2/environment.py", line 1045, in generate yield self.environment.handle_exception(exc_info, True) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/jinja2/environment.py", line 780, in handle_exception reraise(exc_type, exc_value, tb) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/jinja2/_compat.py", line 37, in reraise raise value.with_traceback(tb) File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/synthtool/gcp/templates/java_library/.kokoro/release/publish_javadoc.sh", line 36, in top-level template code NAME={{ metadata['repo']['distribution_name'].split(':')|last }} File "/tmpfs/src/git/autosynth/env/lib/python3.6/site-packages/jinja2/environment.py", line 430, in getattr return getattr(obj, attribute) jinja2.exceptions.UndefinedError: 'dict object' has no attribute 'distribution_name' synthtool > Cleaned up 2 temporary directories. synthtool > Wrote metadata to synth.metadata. Synthesis failed ``` Google internal developers can see the full log [here](https://sponge/f2df4be1-2bc3-402c-9ee3-d298858423e6).
non_process
synthesis failed for java gcloud maven plugin hello autosynth couldn t regenerate java gcloud maven plugin broken heart here s the output from running synth py cloning into working repo switched to branch autosynth running synthtool synthtool executing tmpfs src git autosynth working repo synth py github issue template bug report md github issue template feature request md github issue template support request md github pull request template md github release please yml kokoro build bat kokoro build sh kokoro coerce logs sh kokoro common cfg kokoro continuous common cfg kokoro continuous dependencies cfg kokoro continuous integration cfg kokoro continuous cfg kokoro continuous cfg kokoro continuous osx cfg kokoro continuous win cfg kokoro continuous cfg kokoro continuous lint cfg kokoro continuous propose release cfg kokoro continuous propose release sh kokoro dependencies sh kokoro linkage monitor sh kokoro nightly common cfg kokoro nightly dependencies cfg kokoro nightly integration cfg kokoro nightly cfg kokoro nightly cfg kokoro nightly osx cfg kokoro nightly win cfg kokoro nightly cfg kokoro nightly lint cfg kokoro presubmit clirr cfg kokoro presubmit common cfg kokoro presubmit dependencies cfg kokoro presubmit integration cfg kokoro presubmit cfg kokoro presubmit cfg kokoro presubmit osx cfg kokoro presubmit win cfg kokoro presubmit cfg kokoro presubmit linkage monitor cfg kokoro presubmit lint cfg kokoro release bump snapshot cfg kokoro release bump snapshot sh kokoro release common cfg kokoro release common sh kokoro release drop cfg kokoro release drop sh kokoro release promote cfg kokoro release promote sh kokoro release publish javadoc cfg kokoro release publish javadoc sh traceback most recent call last file home kbuilder pyenv versions lib runpy py line in run module as main main mod spec file home kbuilder pyenv versions lib runpy py line in run code exec code run globals file tmpfs src git autosynth env lib site packages synthtool main py line in main file tmpfs src git autosynth env lib site packages click core py line in call return self main args kwargs file tmpfs src git autosynth env lib site packages click core py line in main rv self invoke ctx file tmpfs src git autosynth env lib site packages click core py line in invoke return ctx invoke self callback ctx params file tmpfs src git autosynth env lib site packages click core py line in invoke return callback args kwargs file tmpfs src git autosynth env lib site packages synthtool main py line in main spec loader exec module synth module type ignore file line in exec module file line in call with frames removed file tmpfs src git autosynth working repo synth py line in templates common templates java library file tmpfs src git autosynth env lib site packages synthtool gcp common py line in java library return self generic library java library kwargs file tmpfs src git autosynth env lib site packages synthtool gcp common py line in generic library result t render kwargs file tmpfs src git autosynth env lib site packages synthtool sources templates py line in render render to path self env template name self dir kwargs file tmpfs src git autosynth env lib site packages synthtool sources templates py line in render to path output dump fh file tmpfs src git autosynth env lib site packages environment py line in dump fp writelines iterable file tmpfs src git autosynth env lib site packages environment py line in next return self next file tmpfs src git autosynth env lib site packages environment py line in generate yield self environment handle exception exc info true file tmpfs src git autosynth env lib site packages environment py line in handle exception reraise exc type exc value tb file tmpfs src git autosynth env lib site packages compat py line in reraise raise value with traceback tb file tmpfs src git autosynth env lib site packages synthtool gcp templates java library kokoro release publish javadoc sh line in top level template code name metadata split last file tmpfs src git autosynth env lib site packages environment py line in getattr return getattr obj attribute exceptions undefinederror dict object has no attribute distribution name synthtool cleaned up temporary directories synthtool wrote metadata to synth metadata synthesis failed google internal developers can see the full log
0
6,698
9,813,960,996
IssuesEvent
2019-06-13 09:10:32
aiidateam/aiida_core
https://api.github.com/repos/aiidateam/aiida_core
opened
Formalization of meaning of exit code ranges
aiida-core 1.x priority/important requires discussion topic/processes type/enhancement
This issue will be a design document to agree on a guideline for the meaning of certain exit code ranges to be used by plugins and potentially even reserve some for internal use in `aiida-core`. The proposal in this message will be updated as the discussion progresses: * `0 - 99` Reserved for internal use by the engine of `aiida-core` * `100 - 199` Reserved for exit codes by scheduler parsers (This only applies to `CalcJobs` and the scheduler parsing is not yet implemented) * `200 - 299` Reserved for input validation errors * `300 - 399` Reserved for unrecoverable errors * `400 - 499` Reserved for recoverable errors
1.0
Formalization of meaning of exit code ranges - This issue will be a design document to agree on a guideline for the meaning of certain exit code ranges to be used by plugins and potentially even reserve some for internal use in `aiida-core`. The proposal in this message will be updated as the discussion progresses: * `0 - 99` Reserved for internal use by the engine of `aiida-core` * `100 - 199` Reserved for exit codes by scheduler parsers (This only applies to `CalcJobs` and the scheduler parsing is not yet implemented) * `200 - 299` Reserved for input validation errors * `300 - 399` Reserved for unrecoverable errors * `400 - 499` Reserved for recoverable errors
process
formalization of meaning of exit code ranges this issue will be a design document to agree on a guideline for the meaning of certain exit code ranges to be used by plugins and potentially even reserve some for internal use in aiida core the proposal in this message will be updated as the discussion progresses reserved for internal use by the engine of aiida core reserved for exit codes by scheduler parsers this only applies to calcjobs and the scheduler parsing is not yet implemented reserved for input validation errors reserved for unrecoverable errors reserved for recoverable errors
1
613,275
19,085,674,518
IssuesEvent
2021-11-29 05:27:41
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
closed
Requesting approach for accessing the service distinct type for service declaration node
Type/Improvement Priority/Blocker Team/CompilerFETools Area/SemanticAPI
**Description:** Need to access service class details through the semantic model when the service has written as with service declaration ```ballerina import ballerina/http; service / on new http:Listener(9090) { // mock implementation } ``` Earlier, we were supposed to check the service is related to the http module by checking its listener module type but with the help of [the new implement](https://github.com/ballerina-platform/ballerina-standard-library/issues/2398), it would be easy for developing other compiler plugins, if we can get details from directly its class. Going through the current serviceNodeDeclaration and accessing its symbol was unable to capture any details regarding the class details to the above example. Did some accessing its parent node using Semantic Model and Syntax tree as well, That time it gives an empty symbol. Could you please let us if there is a way to access class details without reaching listeners type checking ? **Related Issues (optional):** https://github.com/ballerina-platform/openapi-tools/issues/749
1.0
Requesting approach for accessing the service distinct type for service declaration node - **Description:** Need to access service class details through the semantic model when the service has written as with service declaration ```ballerina import ballerina/http; service / on new http:Listener(9090) { // mock implementation } ``` Earlier, we were supposed to check the service is related to the http module by checking its listener module type but with the help of [the new implement](https://github.com/ballerina-platform/ballerina-standard-library/issues/2398), it would be easy for developing other compiler plugins, if we can get details from directly its class. Going through the current serviceNodeDeclaration and accessing its symbol was unable to capture any details regarding the class details to the above example. Did some accessing its parent node using Semantic Model and Syntax tree as well, That time it gives an empty symbol. Could you please let us if there is a way to access class details without reaching listeners type checking ? **Related Issues (optional):** https://github.com/ballerina-platform/openapi-tools/issues/749
non_process
requesting approach for accessing the service distinct type for service declaration node description need to access service class details through the semantic model when the service has written as with service declaration ballerina import ballerina http service on new http listener mock implementation earlier we were supposed to check the service is related to the http module by checking its listener module type but with the help of it would be easy for developing other compiler plugins if we can get details from directly its class going through the current servicenodedeclaration and accessing its symbol was unable to capture any details regarding the class details to the above example did some accessing its parent node using semantic model and syntax tree as well that time it gives an empty symbol could you please let us if there is a way to access class details without reaching listeners type checking related issues optional
0
157,534
19,958,310,086
IssuesEvent
2022-01-28 03:48:19
wangsongc/mavonEditor
https://api.github.com/repos/wangsongc/mavonEditor
opened
CVE-2020-15366 (Medium) detected in ajv-5.5.2.tgz, ajv-4.11.8.tgz
security vulnerability
## CVE-2020-15366 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>ajv-5.5.2.tgz</b>, <b>ajv-4.11.8.tgz</b></p></summary> <p> <details><summary><b>ajv-5.5.2.tgz</b></p></summary> <p>Another JSON Schema Validator</p> <p>Library home page: <a href="https://registry.npmjs.org/ajv/-/ajv-5.5.2.tgz">https://registry.npmjs.org/ajv/-/ajv-5.5.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/ajv/package.json</p> <p> Dependency Hierarchy: - extract-text-webpack-plugin-2.1.2.tgz (Root Library) - schema-utils-0.3.0.tgz - :x: **ajv-5.5.2.tgz** (Vulnerable Library) </details> <details><summary><b>ajv-4.11.8.tgz</b></p></summary> <p>Another JSON Schema Validator</p> <p>Library home page: <a href="https://registry.npmjs.org/ajv/-/ajv-4.11.8.tgz">https://registry.npmjs.org/ajv/-/ajv-4.11.8.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/har-validator/node_modules/ajv/package.json</p> <p> Dependency Hierarchy: - less-2.7.3.tgz (Root Library) - request-2.81.0.tgz - har-validator-4.2.1.tgz - :x: **ajv-4.11.8.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/wangsongc/mavonEditor/commit/fb5f77da2e6a1abdc8035ddac751a729e7652710">fb5f77da2e6a1abdc8035ddac751a729e7652710</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in ajv.validate() in Ajv (aka Another JSON Schema Validator) 6.12.2. A carefully crafted JSON schema could be provided that allows execution of other code by prototype pollution. (While untrusted schemas are recommended against, the worst case of an untrusted schema should be a denial of service, not execution of code.) <p>Publish Date: 2020-07-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15366>CVE-2020-15366</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/ajv-validator/ajv/releases/tag/v6.12.3">https://github.com/ajv-validator/ajv/releases/tag/v6.12.3</a></p> <p>Release Date: 2020-07-15</p> <p>Fix Resolution (ajv): 6.12.3</p> <p>Direct dependency fix Resolution (extract-text-webpack-plugin): 4.0.0-alpha.0</p><p>Fix Resolution (ajv): 6.12.3</p> <p>Direct dependency fix Resolution (less): 3.0.0-RC.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-15366 (Medium) detected in ajv-5.5.2.tgz, ajv-4.11.8.tgz - ## CVE-2020-15366 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>ajv-5.5.2.tgz</b>, <b>ajv-4.11.8.tgz</b></p></summary> <p> <details><summary><b>ajv-5.5.2.tgz</b></p></summary> <p>Another JSON Schema Validator</p> <p>Library home page: <a href="https://registry.npmjs.org/ajv/-/ajv-5.5.2.tgz">https://registry.npmjs.org/ajv/-/ajv-5.5.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/ajv/package.json</p> <p> Dependency Hierarchy: - extract-text-webpack-plugin-2.1.2.tgz (Root Library) - schema-utils-0.3.0.tgz - :x: **ajv-5.5.2.tgz** (Vulnerable Library) </details> <details><summary><b>ajv-4.11.8.tgz</b></p></summary> <p>Another JSON Schema Validator</p> <p>Library home page: <a href="https://registry.npmjs.org/ajv/-/ajv-4.11.8.tgz">https://registry.npmjs.org/ajv/-/ajv-4.11.8.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/har-validator/node_modules/ajv/package.json</p> <p> Dependency Hierarchy: - less-2.7.3.tgz (Root Library) - request-2.81.0.tgz - har-validator-4.2.1.tgz - :x: **ajv-4.11.8.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/wangsongc/mavonEditor/commit/fb5f77da2e6a1abdc8035ddac751a729e7652710">fb5f77da2e6a1abdc8035ddac751a729e7652710</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in ajv.validate() in Ajv (aka Another JSON Schema Validator) 6.12.2. A carefully crafted JSON schema could be provided that allows execution of other code by prototype pollution. (While untrusted schemas are recommended against, the worst case of an untrusted schema should be a denial of service, not execution of code.) <p>Publish Date: 2020-07-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15366>CVE-2020-15366</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/ajv-validator/ajv/releases/tag/v6.12.3">https://github.com/ajv-validator/ajv/releases/tag/v6.12.3</a></p> <p>Release Date: 2020-07-15</p> <p>Fix Resolution (ajv): 6.12.3</p> <p>Direct dependency fix Resolution (extract-text-webpack-plugin): 4.0.0-alpha.0</p><p>Fix Resolution (ajv): 6.12.3</p> <p>Direct dependency fix Resolution (less): 3.0.0-RC.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in ajv tgz ajv tgz cve medium severity vulnerability vulnerable libraries ajv tgz ajv tgz ajv tgz another json schema validator library home page a href path to dependency file package json path to vulnerable library node modules ajv package json dependency hierarchy extract text webpack plugin tgz root library schema utils tgz x ajv tgz vulnerable library ajv tgz another json schema validator library home page a href path to dependency file package json path to vulnerable library node modules har validator node modules ajv package json dependency hierarchy less tgz root library request tgz har validator tgz x ajv tgz vulnerable library found in head commit a href found in base branch master vulnerability details an issue was discovered in ajv validate in ajv aka another json schema validator a carefully crafted json schema could be provided that allows execution of other code by prototype pollution while untrusted schemas are recommended against the worst case of an untrusted schema should be a denial of service not execution of code publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ajv direct dependency fix resolution extract text webpack plugin alpha fix resolution ajv direct dependency fix resolution less rc step up your open source security game with whitesource
0
11,634
5,052,548,446
IssuesEvent
2016-12-21 02:48:53
rust-lang/rust
https://api.github.com/repos/rust-lang/rust
closed
rustbuild: Dist on cross builds produce way too many tarballs
A-rustbuild
The [dist step](https://buildbot.rust-lang.org/builders/nightly-dist-rustc-cross-host-linux/builds/281/steps/distcheck/logs/stdio) on cross builds takes over half an hour, and that's a crazy amount of time for something which should take at most 5 minutes. The [list of tarballs created](https://gist.github.com/alexcrichton/0ef4bdfcb3514e1ff5362394056feb4d) is massively duplicated. Namely we create around 60 source tarballs (!!). Looks like we're also packaging up the standard library dozens of times for the same target. My guess as to what's happening is that by default rustbuild is doing a cross product of (host x target) and doing all the builds. We should skip a massive number of them for distribution and ideally not even put them into the step list at all! Note that this isn't a fire at the moment as everything's just overwriting what was previously created, but this sounds like something that *could* become a fire! Should also be a relatively easy bug to knock out for anyone who wants to become familiar with rustbuild!
1.0
rustbuild: Dist on cross builds produce way too many tarballs - The [dist step](https://buildbot.rust-lang.org/builders/nightly-dist-rustc-cross-host-linux/builds/281/steps/distcheck/logs/stdio) on cross builds takes over half an hour, and that's a crazy amount of time for something which should take at most 5 minutes. The [list of tarballs created](https://gist.github.com/alexcrichton/0ef4bdfcb3514e1ff5362394056feb4d) is massively duplicated. Namely we create around 60 source tarballs (!!). Looks like we're also packaging up the standard library dozens of times for the same target. My guess as to what's happening is that by default rustbuild is doing a cross product of (host x target) and doing all the builds. We should skip a massive number of them for distribution and ideally not even put them into the step list at all! Note that this isn't a fire at the moment as everything's just overwriting what was previously created, but this sounds like something that *could* become a fire! Should also be a relatively easy bug to knock out for anyone who wants to become familiar with rustbuild!
non_process
rustbuild dist on cross builds produce way too many tarballs the on cross builds takes over half an hour and that s a crazy amount of time for something which should take at most minutes the is massively duplicated namely we create around source tarballs looks like we re also packaging up the standard library dozens of times for the same target my guess as to what s happening is that by default rustbuild is doing a cross product of host x target and doing all the builds we should skip a massive number of them for distribution and ideally not even put them into the step list at all note that this isn t a fire at the moment as everything s just overwriting what was previously created but this sounds like something that could become a fire should also be a relatively easy bug to knock out for anyone who wants to become familiar with rustbuild
0
818,607
30,696,028,955
IssuesEvent
2023-07-26 18:40:13
microsoft/devhome
https://api.github.com/repos/microsoft/devhome
closed
Add 'Gaming Development' section in the App Installation flow of 'Machine Configuration'
Issue-Feature Area-Machine-Config Priority-1 Area-Machine-Config-Install
### Suggested new feature or improvement Add a new section to the App Installation flow for gaming development packages to be discoverable by users. ### Scenario When a user goes into the app installation flow, they see a 'Gaming Development' section and are able to find relevant gaming dev packages (e.g. Microsoft GDK). ### Additional details Will update this soon with a list of packages to start with.
1.0
Add 'Gaming Development' section in the App Installation flow of 'Machine Configuration' - ### Suggested new feature or improvement Add a new section to the App Installation flow for gaming development packages to be discoverable by users. ### Scenario When a user goes into the app installation flow, they see a 'Gaming Development' section and are able to find relevant gaming dev packages (e.g. Microsoft GDK). ### Additional details Will update this soon with a list of packages to start with.
non_process
add gaming development section in the app installation flow of machine configuration suggested new feature or improvement add a new section to the app installation flow for gaming development packages to be discoverable by users scenario when a user goes into the app installation flow they see a gaming development section and are able to find relevant gaming dev packages e g microsoft gdk additional details will update this soon with a list of packages to start with
0
288,976
21,729,561,495
IssuesEvent
2022-05-11 10:42:47
Open-Systems-Pharmacology/OSPSuite-R
https://api.github.com/repos/Open-Systems-Pharmacology/OSPSuite-R
opened
Documenting plotting functions using `DataCombined`
subsystem: documentation
Part 2 of #756 Adding examples to the manual; and more detailed examples of workflow in a vignette. - [ ] `plotIndividualTimeProfile()` - [ ] `plotPopulationTimeProfile()`
1.0
Documenting plotting functions using `DataCombined` - Part 2 of #756 Adding examples to the manual; and more detailed examples of workflow in a vignette. - [ ] `plotIndividualTimeProfile()` - [ ] `plotPopulationTimeProfile()`
non_process
documenting plotting functions using datacombined part of adding examples to the manual and more detailed examples of workflow in a vignette plotindividualtimeprofile plotpopulationtimeprofile
0