Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
16,429
31,849,430,595
IssuesEvent
2023-09-14 23:24:55
cabforum/servercert
https://api.github.com/repos/cabforum/servercert
opened
Minor grammatical error in 7.1.2.11.4 Subject Key Identifier
baseline-requirements clean-up
This section contains the following sentence: > For example, CAs may generate the subject key identifier using an algorithm derived from the public key, or may generate a sufficiently-large unique number, such by using a CSPRNG. I think the final clause is either missing an "as" or has an erroneous "such" included, i.e. > For example, CAs may generate the subject key identifier using an algorithm derived from the public key, or may generate a sufficiently-large unique number, such as by using a CSPRNG. OR > For example, CAs may generate the subject key identifier using an algorithm derived from the public key, or may generate a sufficiently-large unique number, by using a CSPRNG.
1.0
Minor grammatical error in 7.1.2.11.4 Subject Key Identifier - This section contains the following sentence: > For example, CAs may generate the subject key identifier using an algorithm derived from the public key, or may generate a sufficiently-large unique number, such by using a CSPRNG. I think the final clause is either missing an "as" or has an erroneous "such" included, i.e. > For example, CAs may generate the subject key identifier using an algorithm derived from the public key, or may generate a sufficiently-large unique number, such as by using a CSPRNG. OR > For example, CAs may generate the subject key identifier using an algorithm derived from the public key, or may generate a sufficiently-large unique number, by using a CSPRNG.
non_process
minor grammatical error in subject key identifier this section contains the following sentence for example cas may generate the subject key identifier using an algorithm derived from the public key or may generate a sufficiently large unique number such by using a csprng i think the final clause is either missing an as or has an erroneous such included i e for example cas may generate the subject key identifier using an algorithm derived from the public key or may generate a sufficiently large unique number such as by using a csprng or for example cas may generate the subject key identifier using an algorithm derived from the public key or may generate a sufficiently large unique number by using a csprng
0
9,099
12,178,381,774
IssuesEvent
2020-04-28 08:54:55
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
Double free detected
high priority module: dataloader module: multiprocessing quansight-nack topic: crash triaged
## 🐛 Bug <!-- A clear and concise description of what the bug is. --> With PyTorch data loader, I am getting an error-> **double-free detected in tache2**. The number of workers was 5 when I got this error. The program is running fine with num_workers 0. cc @ezyang @gchanan @zou3519 @SsnL
1.0
Double free detected - ## 🐛 Bug <!-- A clear and concise description of what the bug is. --> With PyTorch data loader, I am getting an error-> **double-free detected in tache2**. The number of workers was 5 when I got this error. The program is running fine with num_workers 0. cc @ezyang @gchanan @zou3519 @SsnL
process
double free detected 🐛 bug with pytorch data loader i am getting an error double free detected in the number of workers was when i got this error the program is running fine with num workers cc ezyang gchanan ssnl
1
179,286
30,213,276,190
IssuesEvent
2023-07-05 14:04:49
RRZE-Webteam/rrze-lectures
https://api.github.com/repos/RRZE-Webteam/rrze-lectures
closed
format-template für Shortcode?
Design In Diskussion
Dieser Issue ist nur ein Brainstorming/Idee. Ich denke nicht, dass es wirklich notwendig ist, es zu machen. Aber vielleicht doch: Wenn wir jetzt mit den verschiedenen Issues die Ausgabe gut und usable gestalten, wird diese zwangsläufig auch nicht mehr so aussehen wie in Campo/HIS. Könnte es dann aber nicht doch noch Anwender geben, die exakt die Campo-Darstellung wünschen? Wenn ja, wollen wir das unterstützen durch einen Parameter für den Shortcode wie z.B. `template="campo"` ? Wenn das gesetzt würde, würden alle show/hide-Parameter ignoriert werden und die Ausgabe würde dann tatsächlich genauso aussehen wie bei Campo. Macht dass Sinn? Oder wollen wir das bewusst nicht?
1.0
format-template für Shortcode? - Dieser Issue ist nur ein Brainstorming/Idee. Ich denke nicht, dass es wirklich notwendig ist, es zu machen. Aber vielleicht doch: Wenn wir jetzt mit den verschiedenen Issues die Ausgabe gut und usable gestalten, wird diese zwangsläufig auch nicht mehr so aussehen wie in Campo/HIS. Könnte es dann aber nicht doch noch Anwender geben, die exakt die Campo-Darstellung wünschen? Wenn ja, wollen wir das unterstützen durch einen Parameter für den Shortcode wie z.B. `template="campo"` ? Wenn das gesetzt würde, würden alle show/hide-Parameter ignoriert werden und die Ausgabe würde dann tatsächlich genauso aussehen wie bei Campo. Macht dass Sinn? Oder wollen wir das bewusst nicht?
non_process
format template für shortcode dieser issue ist nur ein brainstorming idee ich denke nicht dass es wirklich notwendig ist es zu machen aber vielleicht doch wenn wir jetzt mit den verschiedenen issues die ausgabe gut und usable gestalten wird diese zwangsläufig auch nicht mehr so aussehen wie in campo his könnte es dann aber nicht doch noch anwender geben die exakt die campo darstellung wünschen wenn ja wollen wir das unterstützen durch einen parameter für den shortcode wie z b template campo wenn das gesetzt würde würden alle show hide parameter ignoriert werden und die ausgabe würde dann tatsächlich genauso aussehen wie bei campo macht dass sinn oder wollen wir das bewusst nicht
0
749,382
26,161,786,422
IssuesEvent
2022-12-31 17:09:47
OpenTabletDriver/OpenTabletDriver
https://api.github.com/repos/OpenTabletDriver/OpenTabletDriver
closed
Hawku and Devocub filters feel weird after minimizing OTD into the taskbar
bug help wanted windows/wpf priority:medium
## Description Whenever i minimize OTD into the taskbar the filters start feeling 60hz on polling rate instead of the value i set in the plugin settings, same happens when running the daemon instead of the UX and sadly i cannot try this on windows 10 neither linux since i own a 12th gen intel cpu. Nothing shows up in the console and when testing on osu it shows the right polling rate but it still feels like it is set to 60hz or lower. The only temporary fix i've found is to re-enable the filters and save&apply while not minimizing the OTD and letting the UX run in the background. Do you guys know what might be causing this? Is it maybe a windows update causing this? It started happening a few weeks ago for me and until then it was fine. ## System Information: | Name | Value | | ---------------- | ----- | | Operating System | Windows 11 | | OpenTabletDriver Version | v0.6.0.0 | | Tablet | CTL-472 |
1.0
Hawku and Devocub filters feel weird after minimizing OTD into the taskbar - ## Description Whenever i minimize OTD into the taskbar the filters start feeling 60hz on polling rate instead of the value i set in the plugin settings, same happens when running the daemon instead of the UX and sadly i cannot try this on windows 10 neither linux since i own a 12th gen intel cpu. Nothing shows up in the console and when testing on osu it shows the right polling rate but it still feels like it is set to 60hz or lower. The only temporary fix i've found is to re-enable the filters and save&apply while not minimizing the OTD and letting the UX run in the background. Do you guys know what might be causing this? Is it maybe a windows update causing this? It started happening a few weeks ago for me and until then it was fine. ## System Information: | Name | Value | | ---------------- | ----- | | Operating System | Windows 11 | | OpenTabletDriver Version | v0.6.0.0 | | Tablet | CTL-472 |
non_process
hawku and devocub filters feel weird after minimizing otd into the taskbar description whenever i minimize otd into the taskbar the filters start feeling on polling rate instead of the value i set in the plugin settings same happens when running the daemon instead of the ux and sadly i cannot try this on windows neither linux since i own a gen intel cpu nothing shows up in the console and when testing on osu it shows the right polling rate but it still feels like it is set to or lower the only temporary fix i ve found is to re enable the filters and save apply while not minimizing the otd and letting the ux run in the background do you guys know what might be causing this is it maybe a windows update causing this it started happening a few weeks ago for me and until then it was fine system information name value operating system windows opentabletdriver version tablet ctl
0
14,452
17,533,226,402
IssuesEvent
2021-08-12 01:46:14
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
TIN Mesh Creation processing alg does not correctly apply CRS
Feedback stale Processing Bug Mesh
When I create a new mesh from a point collection, and specify the output CRS, the generated mesh loads but indicates that it has no valid CRS. The sample dataset I used is provided as an attachment to https://github.com/qgis/QGIS/issues/39394 so I won't replicate it here. Here is the workflow I used to create a grid: ![mesh-crs-issue](https://user-images.githubusercontent.com/178003/96136441-b4e9e000-0ef3-11eb-88e0-31f3da593149.gif) After the grid is added to the canvas, it indicates that it has no valid CRS. CC @vcloarec
1.0
TIN Mesh Creation processing alg does not correctly apply CRS - When I create a new mesh from a point collection, and specify the output CRS, the generated mesh loads but indicates that it has no valid CRS. The sample dataset I used is provided as an attachment to https://github.com/qgis/QGIS/issues/39394 so I won't replicate it here. Here is the workflow I used to create a grid: ![mesh-crs-issue](https://user-images.githubusercontent.com/178003/96136441-b4e9e000-0ef3-11eb-88e0-31f3da593149.gif) After the grid is added to the canvas, it indicates that it has no valid CRS. CC @vcloarec
process
tin mesh creation processing alg does not correctly apply crs when i create a new mesh from a point collection and specify the output crs the generated mesh loads but indicates that it has no valid crs the sample dataset i used is provided as an attachment to so i won t replicate it here here is the workflow i used to create a grid after the grid is added to the canvas it indicates that it has no valid crs cc vcloarec
1
18,648
3,398,032,547
IssuesEvent
2015-12-02 00:45:29
Microsoft/TypeScript
https://api.github.com/repos/Microsoft/TypeScript
closed
Incorrect function return type: expected it to match generic type parameter
By Design Question
I may just be abusing the type system here, but there's no warnings or errors, so I have to assume it should work. I'm trying to dispatch a command to a registered handler and return the result, which should be of the type specified by the command. Question is, why is the result always of type object, when the type parameter is clearly a string? Is this a bug or known limitation, and if so, is it a bug that there is no error or warning here? ```typescript interface Command<T> { } class TestCommand implements Command<string> { } function dispatch<T>(command: Command<T>): T { let result: any = "Result of executing the command handler"; return result as T; } let result = dispatch(new TestCommand()); // Why is result of type object? I expected it to be a string. ```
1.0
Incorrect function return type: expected it to match generic type parameter - I may just be abusing the type system here, but there's no warnings or errors, so I have to assume it should work. I'm trying to dispatch a command to a registered handler and return the result, which should be of the type specified by the command. Question is, why is the result always of type object, when the type parameter is clearly a string? Is this a bug or known limitation, and if so, is it a bug that there is no error or warning here? ```typescript interface Command<T> { } class TestCommand implements Command<string> { } function dispatch<T>(command: Command<T>): T { let result: any = "Result of executing the command handler"; return result as T; } let result = dispatch(new TestCommand()); // Why is result of type object? I expected it to be a string. ```
non_process
incorrect function return type expected it to match generic type parameter i may just be abusing the type system here but there s no warnings or errors so i have to assume it should work i m trying to dispatch a command to a registered handler and return the result which should be of the type specified by the command question is why is the result always of type object when the type parameter is clearly a string is this a bug or known limitation and if so is it a bug that there is no error or warning here typescript interface command class testcommand implements command function dispatch command command t let result any result of executing the command handler return result as t let result dispatch new testcommand why is result of type object i expected it to be a string
0
174,392
27,632,596,992
IssuesEvent
2023-03-10 12:04:54
ethereum/solidity
https://api.github.com/repos/ethereum/solidity
closed
Warn if order in override specifier is different than in inheritance list
language design :rage4: closed-due-inactivity stale
Non-breaking part of https://github.com/ethereum/solidity/issues/8354
1.0
Warn if order in override specifier is different than in inheritance list - Non-breaking part of https://github.com/ethereum/solidity/issues/8354
non_process
warn if order in override specifier is different than in inheritance list non breaking part of
0
126,244
26,809,316,877
IssuesEvent
2023-02-01 20:55:43
appsmithorg/appsmith
https://api.github.com/repos/appsmithorg/appsmith
closed
[Bug]: Custom JS Lib return 401 error for anonymous user
Bug High Needs Triaging Mongo BE Coders Pod Integrations Pod Custom JS Libraries
### Is there an existing issue for this? - [X] I have searched the existing issues ### Description https://theappsmith.slack.com/archives/CGBPVEJ5C/p1674729997721399 ### Steps To Reproduce https://theappsmith.slack.com/archives/CGBPVEJ5C/p1674729997721399 ### Public Sample App _No response_ ### Issue video log _No response_ ### Version cloud, self hosted
1.0
[Bug]: Custom JS Lib return 401 error for anonymous user - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Description https://theappsmith.slack.com/archives/CGBPVEJ5C/p1674729997721399 ### Steps To Reproduce https://theappsmith.slack.com/archives/CGBPVEJ5C/p1674729997721399 ### Public Sample App _No response_ ### Issue video log _No response_ ### Version cloud, self hosted
non_process
custom js lib return error for anonymous user is there an existing issue for this i have searched the existing issues description steps to reproduce public sample app no response issue video log no response version cloud self hosted
0
12,531
14,972,332,979
IssuesEvent
2021-01-27 22:39:44
BootBlock/FileSieve
https://api.github.com/repos/BootBlock/FileSieve
opened
Add Move copy mode support to multiple destinations
backend-core processing
Moving files to destination mirrors would probably be more useful than just copying them, but this will require significant core changes that it’ll take a while to implement. Need to tread carefully as a lot of stuff could get broken.
1.0
Add Move copy mode support to multiple destinations - Moving files to destination mirrors would probably be more useful than just copying them, but this will require significant core changes that it’ll take a while to implement. Need to tread carefully as a lot of stuff could get broken.
process
add move copy mode support to multiple destinations moving files to destination mirrors would probably be more useful than just copying them but this will require significant core changes that it’ll take a while to implement need to tread carefully as a lot of stuff could get broken
1
19,150
25,225,841,803
IssuesEvent
2022-11-14 15:55:19
maticnetwork/miden
https://api.github.com/repos/maticnetwork/miden
closed
Memory pre-loading
instruction set processor v0.3
The following issue came up during discussions related to the implementation of the number theoretic transform (NTT). In many situations, it is required to have arrays of large sizes loaded into memory before the actual program starts running. This is the situation for example when doing NTT where a large array containing roots of unity is needed during the iterations of the algorithm. Clearly these constants should reside in memory as they are, most of the time, used several times during the execution of any useful program. The question that poses itself then is how to do this in the most efficient+secure way. Some ideas that came up during the discussion: 1. One might augment the current mechanism used to load advice inputs in order to target the above situations in an efficient manner. 2. The memory trace can be pre-populated with the data by the prover before the start of the execution. This should be better, I believe, than the previous proposal but it is not immediately clear, to me at least, how we can insure soundness with this modification while still using the methodology we currently use for insuring soundness i.e. without any additional permutation-checks machinery.
1.0
Memory pre-loading - The following issue came up during discussions related to the implementation of the number theoretic transform (NTT). In many situations, it is required to have arrays of large sizes loaded into memory before the actual program starts running. This is the situation for example when doing NTT where a large array containing roots of unity is needed during the iterations of the algorithm. Clearly these constants should reside in memory as they are, most of the time, used several times during the execution of any useful program. The question that poses itself then is how to do this in the most efficient+secure way. Some ideas that came up during the discussion: 1. One might augment the current mechanism used to load advice inputs in order to target the above situations in an efficient manner. 2. The memory trace can be pre-populated with the data by the prover before the start of the execution. This should be better, I believe, than the previous proposal but it is not immediately clear, to me at least, how we can insure soundness with this modification while still using the methodology we currently use for insuring soundness i.e. without any additional permutation-checks machinery.
process
memory pre loading the following issue came up during discussions related to the implementation of the number theoretic transform ntt in many situations it is required to have arrays of large sizes loaded into memory before the actual program starts running this is the situation for example when doing ntt where a large array containing roots of unity is needed during the iterations of the algorithm clearly these constants should reside in memory as they are most of the time used several times during the execution of any useful program the question that poses itself then is how to do this in the most efficient secure way some ideas that came up during the discussion one might augment the current mechanism used to load advice inputs in order to target the above situations in an efficient manner the memory trace can be pre populated with the data by the prover before the start of the execution this should be better i believe than the previous proposal but it is not immediately clear to me at least how we can insure soundness with this modification while still using the methodology we currently use for insuring soundness i e without any additional permutation checks machinery
1
93,167
3,886,536,936
IssuesEvent
2016-04-14 01:40:02
nnnick/Chart.js
https://api.github.com/repos/nnnick/Chart.js
closed
generateLegend() not generating a legend correctly
Category: Bug Priority: p1 Version: 2.x
Hi, Don't know if this has been fixed or not (I couldn't see it after doing a search) but currently the generateLegend function is producing strange results: ![capture](https://cloud.githubusercontent.com/assets/9978497/14494835/c51fd47e-0184-11e6-856c-c3defdcea17b.JPG) As you can see the label has the background color applied to it, when really we want it to be applied to a span next to it. I've looked at the chart js code and realised that it's to do with the way the html is created in the legendCallback. The span is around the text as well when surely it should be next to it to create the colour block? Like so: ![capture jpg2](https://cloud.githubusercontent.com/assets/9978497/14494987/7a81e802-0185-11e6-8c34-5ecc1d02874f.JPG) In order to fix this I changed the legendCallback function on line 105 to: ``` legendCallback: function(chart) { var text = []; text.push('<ul class="' + chart.id + '-legend">'); for (var i = 0; i < chart.data.datasets.length; i++) { text.push('<li><span style="background-color:' + chart.data.datasets[i].backgroundColor + '"></span>'); if (chart.data.datasets[i].label) { text.push(chart.data.datasets[i].label); } text.push('</li>'); } text.push('</ul>'); return text.join(""); } ```
1.0
generateLegend() not generating a legend correctly - Hi, Don't know if this has been fixed or not (I couldn't see it after doing a search) but currently the generateLegend function is producing strange results: ![capture](https://cloud.githubusercontent.com/assets/9978497/14494835/c51fd47e-0184-11e6-856c-c3defdcea17b.JPG) As you can see the label has the background color applied to it, when really we want it to be applied to a span next to it. I've looked at the chart js code and realised that it's to do with the way the html is created in the legendCallback. The span is around the text as well when surely it should be next to it to create the colour block? Like so: ![capture jpg2](https://cloud.githubusercontent.com/assets/9978497/14494987/7a81e802-0185-11e6-8c34-5ecc1d02874f.JPG) In order to fix this I changed the legendCallback function on line 105 to: ``` legendCallback: function(chart) { var text = []; text.push('<ul class="' + chart.id + '-legend">'); for (var i = 0; i < chart.data.datasets.length; i++) { text.push('<li><span style="background-color:' + chart.data.datasets[i].backgroundColor + '"></span>'); if (chart.data.datasets[i].label) { text.push(chart.data.datasets[i].label); } text.push('</li>'); } text.push('</ul>'); return text.join(""); } ```
non_process
generatelegend not generating a legend correctly hi don t know if this has been fixed or not i couldn t see it after doing a search but currently the generatelegend function is producing strange results as you can see the label has the background color applied to it when really we want it to be applied to a span next to it i ve looked at the chart js code and realised that it s to do with the way the html is created in the legendcallback the span is around the text as well when surely it should be next to it to create the colour block like so in order to fix this i changed the legendcallback function on line to legendcallback function chart var text text push for var i i chart data datasets length i text push if chart data datasets label text push chart data datasets label text push text push return text join
0
15,591
19,717,131,511
IssuesEvent
2022-01-13 12:08:55
googleapis/python-pubsub
https://api.github.com/repos/googleapis/python-pubsub
closed
Your .repo-metadata.json file has a problem 🤒
api: pubsub type: process repo-metadata: lint
You have a problem with your .repo-metadata.json file: Result of scan 📈: * library_type must be equal to one of the allowed values in .repo-metadata.json ☝️ Once you address these problems, you can close this issue. ### Need help? * [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field. * [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**. * Reach out to **go/github-automation** if you have any questions.
1.0
Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file: Result of scan 📈: * library_type must be equal to one of the allowed values in .repo-metadata.json ☝️ Once you address these problems, you can close this issue. ### Need help? * [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field. * [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**. * Reach out to **go/github-automation** if you have any questions.
process
your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 library type must be equal to one of the allowed values in repo metadata json ☝️ once you address these problems you can close this issue need help lists valid options for each field for grpc libraries api shortname should match the subdomain of an api s hostname reach out to go github automation if you have any questions
1
13,306
8,409,370,039
IssuesEvent
2018-10-12 07:03:55
virtual-labs/problem-solving-iiith
https://api.github.com/repos/virtual-labs/problem-solving-iiith
closed
QA_Problem-Solving_Manual_Beauty-of-Numbers_Spelling-Mistakes
Category :Usability Developed by:IIIT Hyd Open-edx-Issue Severity :S2 Status :Resolved
Defect Description : While testing this lab, found spelling mistakes in the Manual page of Beauty of Numbers experiment. Actual Result : While testing this lab, found spelling mistakes in the Manual page of Beauty of Numbers experiment. Please refer to the screen shot provided in the attachments. Environment : OS: Windows 7, Ubuntu-16.04, Centos-6 Browsers: Firefox-42.0,Chrome-47.0,chromium-45.0 Bandwidth : 100Mbps Hardware Configuration:8GBRAM, Processor:i5 Attachments: ![qa_openedx_ps_i18](https://cloud.githubusercontent.com/assets/13479177/25838907/f12ba388-34b2-11e7-9f54-a87542065c7e.png)
True
QA_Problem-Solving_Manual_Beauty-of-Numbers_Spelling-Mistakes - Defect Description : While testing this lab, found spelling mistakes in the Manual page of Beauty of Numbers experiment. Actual Result : While testing this lab, found spelling mistakes in the Manual page of Beauty of Numbers experiment. Please refer to the screen shot provided in the attachments. Environment : OS: Windows 7, Ubuntu-16.04, Centos-6 Browsers: Firefox-42.0,Chrome-47.0,chromium-45.0 Bandwidth : 100Mbps Hardware Configuration:8GBRAM, Processor:i5 Attachments: ![qa_openedx_ps_i18](https://cloud.githubusercontent.com/assets/13479177/25838907/f12ba388-34b2-11e7-9f54-a87542065c7e.png)
non_process
qa problem solving manual beauty of numbers spelling mistakes defect description while testing this lab found spelling mistakes in the manual page of beauty of numbers experiment actual result while testing this lab found spelling mistakes in the manual page of beauty of numbers experiment please refer to the screen shot provided in the attachments environment os windows ubuntu centos browsers firefox chrome chromium bandwidth hardware configuration processor attachments
0
3,836
6,802,432,608
IssuesEvent
2017-11-02 20:11:19
gratipay/inside.gratipay.com
https://api.github.com/repos/gratipay/inside.gratipay.com
closed
we need a work contract
Compliance Governance & Process
Redux of https://github.com/gratipay/gratipay.com/issues/1443. We need a contract in place with any employees or contractors to whom we give permission to manage sensitive data.
1.0
we need a work contract - Redux of https://github.com/gratipay/gratipay.com/issues/1443. We need a contract in place with any employees or contractors to whom we give permission to manage sensitive data.
process
we need a work contract redux of we need a contract in place with any employees or contractors to whom we give permission to manage sensitive data
1
47,327
6,046,269,640
IssuesEvent
2017-06-12 11:36:35
hnguyenworkstation/bottlechat
https://api.github.com/repos/hnguyenworkstation/bottlechat
closed
First-time User -- Lead them to a layout to setup their information
design feature
- First-time User -- Lead them to a layout to setup their information - Used for first-time user that logged in using apis like Facebook, Google - Collect their emails and provide them a layout to them to setup: - Passwords - Names ... etc...
1.0
First-time User -- Lead them to a layout to setup their information - - First-time User -- Lead them to a layout to setup their information - Used for first-time user that logged in using apis like Facebook, Google - Collect their emails and provide them a layout to them to setup: - Passwords - Names ... etc...
non_process
first time user lead them to a layout to setup their information first time user lead them to a layout to setup their information used for first time user that logged in using apis like facebook google collect their emails and provide them a layout to them to setup passwords names etc
0
8,147
11,354,715,095
IssuesEvent
2020-01-24 18:17:07
googleapis/java-webrisk
https://api.github.com/repos/googleapis/java-webrisk
closed
Promote to Beta
type: process
Package name: **google-cloud-webrisk** Current release: **alpha** Proposed release: **beta** ## Instructions Check the lists below, adding tests / documentation as required. Once all the "required" boxes are ticked, please create a release and close this issue. ## Required - [x] Server API is beta or GA - [x] Service API is public - [x] Client surface is mostly stable (no known issues that could significantly change the surface) - [ ] All manual types and methods have comment documentation - [ ] Package name is idiomatic for the platform - [ ] At least one integration/smoke test is defined and passing - [ ] Central GitHub README lists and points to the per-API README - [ ] Per-API README links to product page on cloud.google.com - [ ] Manual code has been reviewed for API stability by repo owner ## Optional - [ ] Most common / important scenarios have descriptive samples - [ ] Public manual methods have at least one usage sample each (excluding overloads) - [ ] Per-API README includes a full description of the API - [ ] Per-API README contains at least one “getting started” sample using the most common API scenario - [ ] Manual code has been reviewed by API producer - [ ] Manual code has been reviewed by a DPE responsible for samples - [ ] 'Client LIbraries' page is added to the product documentation in 'APIs & Reference' section of the product's documentation on Cloud Site
1.0
Promote to Beta - Package name: **google-cloud-webrisk** Current release: **alpha** Proposed release: **beta** ## Instructions Check the lists below, adding tests / documentation as required. Once all the "required" boxes are ticked, please create a release and close this issue. ## Required - [x] Server API is beta or GA - [x] Service API is public - [x] Client surface is mostly stable (no known issues that could significantly change the surface) - [ ] All manual types and methods have comment documentation - [ ] Package name is idiomatic for the platform - [ ] At least one integration/smoke test is defined and passing - [ ] Central GitHub README lists and points to the per-API README - [ ] Per-API README links to product page on cloud.google.com - [ ] Manual code has been reviewed for API stability by repo owner ## Optional - [ ] Most common / important scenarios have descriptive samples - [ ] Public manual methods have at least one usage sample each (excluding overloads) - [ ] Per-API README includes a full description of the API - [ ] Per-API README contains at least one “getting started” sample using the most common API scenario - [ ] Manual code has been reviewed by API producer - [ ] Manual code has been reviewed by a DPE responsible for samples - [ ] 'Client LIbraries' page is added to the product documentation in 'APIs & Reference' section of the product's documentation on Cloud Site
process
promote to beta package name google cloud webrisk current release alpha proposed release beta instructions check the lists below adding tests documentation as required once all the required boxes are ticked please create a release and close this issue required server api is beta or ga service api is public client surface is mostly stable no known issues that could significantly change the surface all manual types and methods have comment documentation package name is idiomatic for the platform at least one integration smoke test is defined and passing central github readme lists and points to the per api readme per api readme links to product page on cloud google com manual code has been reviewed for api stability by repo owner optional most common important scenarios have descriptive samples public manual methods have at least one usage sample each excluding overloads per api readme includes a full description of the api per api readme contains at least one “getting started” sample using the most common api scenario manual code has been reviewed by api producer manual code has been reviewed by a dpe responsible for samples client libraries page is added to the product documentation in apis reference section of the product s documentation on cloud site
1
21,195
28,213,914,183
IssuesEvent
2023-04-05 07:30:13
WICG/admin
https://api.github.com/repos/WICG/admin
closed
A unique identifier for W3C specifications
Incubation Process
It'd be nice if the W3C specifications had a unique ID that we could use to query APIs similarly to what is done with other standardization instances specifications. [IETF Internet-Drafts & RFC](https://datatracker.ietf.org/doc/html/rfc2648) have an ID in the form of `urn:ietf:id:draft-blank-ietf-bimi-02` and `urn:ietf:rfc:9114`, and can be used with the [IETF API](https://datatracker.ietf.org/api/). Similarly, there exists the same for [ETSI specifications](https://datatracker.ietf.org/doc/html/rfc2648), e.g `urn:etsi:102818` and [XSF specifications](https://datatracker.ietf.org/doc/html/rfc4854) (with spec versioning) e.g `urn:xmpp:feature-fallback:0`.
1.0
A unique identifier for W3C specifications - It'd be nice if the W3C specifications had a unique ID that we could use to query APIs similarly to what is done with other standardization instances specifications. [IETF Internet-Drafts & RFC](https://datatracker.ietf.org/doc/html/rfc2648) have an ID in the form of `urn:ietf:id:draft-blank-ietf-bimi-02` and `urn:ietf:rfc:9114`, and can be used with the [IETF API](https://datatracker.ietf.org/api/). Similarly, there exists the same for [ETSI specifications](https://datatracker.ietf.org/doc/html/rfc2648), e.g `urn:etsi:102818` and [XSF specifications](https://datatracker.ietf.org/doc/html/rfc4854) (with spec versioning) e.g `urn:xmpp:feature-fallback:0`.
process
a unique identifier for specifications it d be nice if the specifications had a unique id that we could use to query apis similarly to what is done with other standardization instances specifications have an id in the form of urn ietf id draft blank ietf bimi and urn ietf rfc and can be used with the similarly there exists the same for e g urn etsi and with spec versioning e g urn xmpp feature fallback
1
2,273
5,103,843,014
IssuesEvent
2017-01-04 22:43:06
Jarvvski/CavTools
https://api.github.com/repos/Jarvvski/CavTools
closed
BF1 Origins on Enlisment App
Process Flow Request
<h2>Problem</h2><br />Could you please look into having a box to input applicant's Origins name on the enlistment form?<br /><hr><h2>Reason</h2><br />Makes it easier to track applicants<br><br>-Specialist Masse.P
1.0
BF1 Origins on Enlisment App - <h2>Problem</h2><br />Could you please look into having a box to input applicant's Origins name on the enlistment form?<br /><hr><h2>Reason</h2><br />Makes it easier to track applicants<br><br>-Specialist Masse.P
process
origins on enlisment app problem could you please look into having a box to input applicant s origins name on the enlistment form reason makes it easier to track applicants specialist masse p
1
12,892
9,996,704,884
IssuesEvent
2019-07-12 00:51:46
APSIMInitiative/ApsimX
https://api.github.com/repos/APSIMInitiative/ApsimX
closed
Unnecessary serialization of simulation properties
interface/infrastructure refactor
Certain properties of simulation objects (Descriptors and Services) are serialised when they do not need to be.
1.0
Unnecessary serialization of simulation properties - Certain properties of simulation objects (Descriptors and Services) are serialised when they do not need to be.
non_process
unnecessary serialization of simulation properties certain properties of simulation objects descriptors and services are serialised when they do not need to be
0
21,324
28,960,678,841
IssuesEvent
2023-05-10 02:01:29
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
Bazel Deps not linking on `cc_library` but links on `cc_binary`
P3 type: support / not a bug (process) team-Rules-CPP stale
I am attempting to include a library called pytorch in my project ``` cc_library( name = "torch_cpu", srcs = [ "lib/libtorch.so", "lib/libtorch_cpu.so", "lib/libc10.so", "lib/libcaffe2_detectron_ops.so", ], #srcs = glob(["lib/*.so"]), linkopts = [ "-ltorch", "-ltorch_cpu", "-lcaffe2_detectron_ops", "-lc10", ], hdrs = glob(["include/**/*.h"]), includes = [ "include", "include/torch/csrc/api/include", "include/THC", "include/TH", ], copts = ["-D_GLIBCXX_USE_CXX11_ABI=0"], visibility = ["//visibility:public"] ) ``` When I am building a binary as such: ``` cc_binary( name = "bin", srcs = ["bin.cpp"], linkshared = True, deps = [ "@libtorch//:torch_cpu", ], ) ``` It generates a `libbin.so` and i see libtorch is linked when i do ldd on it ``` ldd bazel-bin/libbin.so linux-vdso.so.1 (0x00007fff84792000) libtorch.so => /home/raaj/.cache/bazel/_bazel_raaj/be6e4ec0d0de66f5dc2830250d02e4c6/external/pytorch_cpu/libtorch/lib/libtorch.so (0x00007fd987bcc000) ... ``` But when i am build a library: ``` cc_library( name = "lib", srcs = ["lib.cpp"], hdrs = ["lib.hpp"], deps = [ "@libtorch//:torch_cpu", ], ) ``` It generates a `liblib.so` and i do not see it linked ``` ldd bazel-bin/liblib.so linux-vdso.so.1 (0x00007ffe5e887000) libstdc++.so.6 => /usr/lib/x86_64-linux-gnu/libstdc++.so.6 (0x00007f2715b4d000) libm.so.6 => /lib/x86_64-linux-gnu/libm.so.6 (0x00007f27157af000) libgcc_s.so.1 => /lib/x86_64-linux-gnu/libgcc_s.so.1 (0x00007f2715597000) libc.so.6 => /lib/x86_64-linux-gnu/libc.so.6 (0x00007f27151a6000) /lib64/ld-linux-x86-64.so.2 (0x00007f2715f5a000) ``` Why is this?
1.0
Bazel Deps not linking on `cc_library` but links on `cc_binary` - I am attempting to include a library called pytorch in my project ``` cc_library( name = "torch_cpu", srcs = [ "lib/libtorch.so", "lib/libtorch_cpu.so", "lib/libc10.so", "lib/libcaffe2_detectron_ops.so", ], #srcs = glob(["lib/*.so"]), linkopts = [ "-ltorch", "-ltorch_cpu", "-lcaffe2_detectron_ops", "-lc10", ], hdrs = glob(["include/**/*.h"]), includes = [ "include", "include/torch/csrc/api/include", "include/THC", "include/TH", ], copts = ["-D_GLIBCXX_USE_CXX11_ABI=0"], visibility = ["//visibility:public"] ) ``` When I am building a binary as such: ``` cc_binary( name = "bin", srcs = ["bin.cpp"], linkshared = True, deps = [ "@libtorch//:torch_cpu", ], ) ``` It generates a `libbin.so` and i see libtorch is linked when i do ldd on it ``` ldd bazel-bin/libbin.so linux-vdso.so.1 (0x00007fff84792000) libtorch.so => /home/raaj/.cache/bazel/_bazel_raaj/be6e4ec0d0de66f5dc2830250d02e4c6/external/pytorch_cpu/libtorch/lib/libtorch.so (0x00007fd987bcc000) ... ``` But when i am build a library: ``` cc_library( name = "lib", srcs = ["lib.cpp"], hdrs = ["lib.hpp"], deps = [ "@libtorch//:torch_cpu", ], ) ``` It generates a `liblib.so` and i do not see it linked ``` ldd bazel-bin/liblib.so linux-vdso.so.1 (0x00007ffe5e887000) libstdc++.so.6 => /usr/lib/x86_64-linux-gnu/libstdc++.so.6 (0x00007f2715b4d000) libm.so.6 => /lib/x86_64-linux-gnu/libm.so.6 (0x00007f27157af000) libgcc_s.so.1 => /lib/x86_64-linux-gnu/libgcc_s.so.1 (0x00007f2715597000) libc.so.6 => /lib/x86_64-linux-gnu/libc.so.6 (0x00007f27151a6000) /lib64/ld-linux-x86-64.so.2 (0x00007f2715f5a000) ``` Why is this?
process
bazel deps not linking on cc library but links on cc binary i am attempting to include a library called pytorch in my project cc library name torch cpu srcs lib libtorch so lib libtorch cpu so lib so lib detectron ops so srcs glob linkopts ltorch ltorch cpu detectron ops hdrs glob includes include include torch csrc api include include thc include th copts visibility when i am building a binary as such cc binary name bin srcs linkshared true deps libtorch torch cpu it generates a libbin so and i see libtorch is linked when i do ldd on it ldd bazel bin libbin so linux vdso so libtorch so home raaj cache bazel bazel raaj external pytorch cpu libtorch lib libtorch so but when i am build a library cc library name lib srcs hdrs deps libtorch torch cpu it generates a liblib so and i do not see it linked ldd bazel bin liblib so linux vdso so libstdc so usr lib linux gnu libstdc so libm so lib linux gnu libm so libgcc s so lib linux gnu libgcc s so libc so lib linux gnu libc so ld linux so why is this
1
17,496
23,305,508,035
IssuesEvent
2022-08-07 23:50:05
lynnandtonic/nestflix.fun
https://api.github.com/repos/lynnandtonic/nestflix.fun
closed
Add Sports Shouting from "30 Rock"
suggested title in process
Please add as much of the following info as you can: Title: Sports Shouting Type (film/tv show): TV show Film or show in which it appears: **30 Rock** Is the parent film/show streaming anywhere? Yes About when in the parent film/show does it appear? S04E05 Actual footage of the film/show can be seen (yes/no)? Yes
1.0
Add Sports Shouting from "30 Rock" - Please add as much of the following info as you can: Title: Sports Shouting Type (film/tv show): TV show Film or show in which it appears: **30 Rock** Is the parent film/show streaming anywhere? Yes About when in the parent film/show does it appear? S04E05 Actual footage of the film/show can be seen (yes/no)? Yes
process
add sports shouting from rock please add as much of the following info as you can title sports shouting type film tv show tv show film or show in which it appears rock is the parent film show streaming anywhere yes about when in the parent film show does it appear actual footage of the film show can be seen yes no yes
1
186,290
14,394,659,959
IssuesEvent
2020-12-03 01:49:23
github-vet/rangeclosure-findings
https://api.github.com/repos/github-vet/rangeclosure-findings
closed
logosapollos/ego: src/pkg/image/jpeg/dct_test.go; 14 LoC
fresh small test
Found a possible issue in [logosapollos/ego](https://www.github.com/logosapollos/ego) at [src/pkg/image/jpeg/dct_test.go](https://github.com/logosapollos/ego/blob/61fab933d8850b59a60a88ef3456f6a7470e9d02/src/pkg/image/jpeg/dct_test.go#L53-L66) The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements which capture loop variables. [Click here to see the code in its original context.](https://github.com/logosapollos/ego/blob/61fab933d8850b59a60a88ef3456f6a7470e9d02/src/pkg/image/jpeg/dct_test.go#L53-L66) <details> <summary>Click here to show the 14 line(s) of Go which triggered the analyzer.</summary> ```go for i, b := range blocks { got, want := b, b for j := range got { got[j] = (got[j] - 128) * 8 } slowFDCT(&got) slowIDCT(&got) for j := range got { got[j] = got[j]/8 + 128 } if differ(&got, &want) { t.Errorf("i=%d: IDCT(FDCT)\nsrc\n%s\ngot\n%s\nwant\n%s\n", i, &b, &got, &want) } } ``` Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to b at line 64 may start a goroutine </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 61fab933d8850b59a60a88ef3456f6a7470e9d02
1.0
logosapollos/ego: src/pkg/image/jpeg/dct_test.go; 14 LoC - Found a possible issue in [logosapollos/ego](https://www.github.com/logosapollos/ego) at [src/pkg/image/jpeg/dct_test.go](https://github.com/logosapollos/ego/blob/61fab933d8850b59a60a88ef3456f6a7470e9d02/src/pkg/image/jpeg/dct_test.go#L53-L66) The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements which capture loop variables. [Click here to see the code in its original context.](https://github.com/logosapollos/ego/blob/61fab933d8850b59a60a88ef3456f6a7470e9d02/src/pkg/image/jpeg/dct_test.go#L53-L66) <details> <summary>Click here to show the 14 line(s) of Go which triggered the analyzer.</summary> ```go for i, b := range blocks { got, want := b, b for j := range got { got[j] = (got[j] - 128) * 8 } slowFDCT(&got) slowIDCT(&got) for j := range got { got[j] = got[j]/8 + 128 } if differ(&got, &want) { t.Errorf("i=%d: IDCT(FDCT)\nsrc\n%s\ngot\n%s\nwant\n%s\n", i, &b, &got, &want) } } ``` Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to b at line 64 may start a goroutine </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 61fab933d8850b59a60a88ef3456f6a7470e9d02
non_process
logosapollos ego src pkg image jpeg dct test go loc found a possible issue in at the below snippet of go code triggered static analysis which searches for goroutines and or defer statements which capture loop variables click here to show the line s of go which triggered the analyzer go for i b range blocks got want b b for j range got got got slowfdct got slowidct got for j range got got got if differ got want t errorf i d idct fdct nsrc n s ngot n s nwant n s n i b got want below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to b at line may start a goroutine leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
0
18,327
24,445,549,113
IssuesEvent
2022-10-06 17:38:15
MPMG-DCC-UFMG/C01
https://api.github.com/repos/MPMG-DCC-UFMG/C01
closed
Problemas ao deletar parâmetros opcionais
[1] Bug [2] Alta Prioridade [0] Desenvolvimento [3] Processamento Dinâmico
## Comportamento Esperado Espera-se que parâmetros opcionais instanciados possam ser removidos quando desejamos retirá-lo ao editarmos um coletor. ## Comportamento Atual Quando editamos um coletor, descartando um parâmetro opcional de um passo dinâmico qualquer, o sistema não salva essa alteração: o parâmetro continua na configuração dos passos. ## Passos para reproduzir o erro 1. Crie um coletor, utilizando algum passo que contenha um parâmetro opcional e instancie-o. Um exemplo de um coletor com essas características está descrito na seção seguinte. Abaixo, o JSON de passos referente a ele: note que há o uso do parâmetro opcional `número de xpaths` do passo `Localizar elementos`. ``` { "step": "root", "depth": 0, "children": [ { "step": "espere", "depth": 1, "arguments": { "segundos": "2" } }, { "step": "para_cada", "depth": 1, "iterator": "ano", "children": [ { "step": "selecione", "depth": 2, "arguments": { "xpath": "'//*[@id=\"NUM_ANO_REFERENCIA\"]'", "opcao": "ano" } }, { "step": "clique", "depth": 2, "arguments": { "elemento": "'//*[@id=\"formDados\"]/fieldset/div/div[7]/input'" } }, { "step": "espere", "depth": 2, "arguments": { "segundos": "3" } }, { "step": "para_cada", "depth": 2, "iterator": "pagina", "children": [ { "step": "espere", "depth": 3, "arguments": { "segundos": "3" } }, { "step": "salva_pagina", "depth": 3, "arguments": {} } ], "iterable": { "call": { "step": "localiza_elementos", "arguments": { "xpath": "'/html/body/div/div/div/div/div/div[2]/center/a[*]'", "numero_xpaths": "3" } } } } ], "iterable": { "call": { "step": "opcoes", "arguments": { "xpath": "'//*[@id=\"NUM_ANO_REFERENCIA\"]'" } } } } ] } ``` 2. Salve o coletor. 3. Edite esse coletor, retirando o parâmetro opcional `número de xpaths`. 4. Ao salvar essa edição, verificamos que esse parâmetro opcional que descartamos continua na configuração. ## Especificações da Coleta Exemplo: ``` { "source_name": "Licita\u00e7\u00f5es Itabirito", "base_url": "http://api.conectbr.com.br/Licitacao/Busca/?token=XEOk3i50lCa5IcpiNG7vXQ==", "obey_robots": false, "data_path": "/home/loui/coletas/licitacoes_itabirito", "request_type": "GET", "form_request_type": "POST", "antiblock_download_delay": 2, "antiblock_autothrottle_enabled": false, "antiblock_autothrottle_start_delay": 2, "antiblock_autothrottle_max_delay": 10, "antiblock_ip_rotation_enabled": false, "antiblock_ip_rotation_type": "tor", "antiblock_max_reqs_per_ip": 10, "antiblock_max_reuse_rounds": 10, "antiblock_proxy_list": "", "antiblock_user_agent_rotation_enabled": false, "antiblock_reqs_per_user_agent": 100, "antiblock_user_agents_list": "", "antiblock_insert_cookies_enabled": false, "antiblock_cookies_list": "", "captcha": "none", "has_webdriver": false, "webdriver_path": null, "img_xpath": null, "sound_xpath": null, "dynamic_processing": true, "explore_links": true, "link_extractor_max_depth": 1, "link_extractor_allow_url": "^(http:\\/\\/api.conectbr.com.br\\/Licitacao\\/AbreArquivo\\/.*)", "link_extractor_allow_domains": null, "link_extractor_tags": null, "link_extractor_attrs": null, "link_extractor_check_type": false, "link_extractor_process_value": "", "download_files": true, "download_files_allow_url": "^(http:\\/\\/api.conectbr.com.br\\/Licitacao\\/AbreArquivo\\/.*)", "download_files_allow_extensions": "pdf,doc", "download_files_allow_domains": null, "download_files_tags": null, "download_files_attrs": null, "download_files_process_value": "", "download_files_check_large_content": true, "download_imgs": false, "steps": "{\"step\":\"root\",\"depth\":0,\"children\":[{\"step\":\"espere\",\"depth\":1,\"arguments\":{\"segundos\":\"2\"}},{\"step\":\"para_cada\",\"depth\":1,\"iterator\":\"ano\",\"children\":[{\"step\":\"selecione\",\"depth\":2,\"arguments\":{\"xpath\":\"'//*[@id=\\\"NUM_ANO_REFERENCIA\\\"]'\",\"opcao\":\"ano\"}},{\"step\":\"clique\",\"depth\":2,\"arguments\":{\"elemento\":\"'//*[@id=\\\"formDados\\\"]/fieldset/div/div[7]/input'\"}},{\"step\":\"espere\",\"depth\":2,\"arguments\":{\"segundos\":\"3\"}},{\"step\":\"para_cada\",\"depth\":2,\"iterator\":\"pagina\",\"children\":[{\"step\":\"espere\",\"depth\":3,\"arguments\":{\"segundos\":\"3\"}},{\"step\":\"salva_pagina\",\"depth\":3,\"arguments\":{}}],\"iterable\":{\"call\":{\"step\":\"localiza_elementos\",\"arguments\":{\"xpath\":\"'/html/body/div/div/div/div/div/div[2]/center/a[*]'\",\"numero_xpaths\":\"3\"}}}}],\"iterable\":{\"call\":{\"step\":\"opcoes\",\"arguments\":{\"xpath\":\"'//*[@id=\\\"NUM_ANO_REFERENCIA\\\"]'\"}}}}]}", "encoding_detection_method": 1, "templated_url_parameter_handlers": [], "static_form_parameter_handlers": [], "templated_url_response_handlers": [], "static_form_response_handlers": [], "crawler_id": 3, "instance_id": "164202556313372" } ``` ## Sistema (caso necessário) - Problema encontrado na branch `master`, provavelmente presente no sistema distribuído e na versão das máquinas do MP.
1.0
Problemas ao deletar parâmetros opcionais - ## Comportamento Esperado Espera-se que parâmetros opcionais instanciados possam ser removidos quando desejamos retirá-lo ao editarmos um coletor. ## Comportamento Atual Quando editamos um coletor, descartando um parâmetro opcional de um passo dinâmico qualquer, o sistema não salva essa alteração: o parâmetro continua na configuração dos passos. ## Passos para reproduzir o erro 1. Crie um coletor, utilizando algum passo que contenha um parâmetro opcional e instancie-o. Um exemplo de um coletor com essas características está descrito na seção seguinte. Abaixo, o JSON de passos referente a ele: note que há o uso do parâmetro opcional `número de xpaths` do passo `Localizar elementos`. ``` { "step": "root", "depth": 0, "children": [ { "step": "espere", "depth": 1, "arguments": { "segundos": "2" } }, { "step": "para_cada", "depth": 1, "iterator": "ano", "children": [ { "step": "selecione", "depth": 2, "arguments": { "xpath": "'//*[@id=\"NUM_ANO_REFERENCIA\"]'", "opcao": "ano" } }, { "step": "clique", "depth": 2, "arguments": { "elemento": "'//*[@id=\"formDados\"]/fieldset/div/div[7]/input'" } }, { "step": "espere", "depth": 2, "arguments": { "segundos": "3" } }, { "step": "para_cada", "depth": 2, "iterator": "pagina", "children": [ { "step": "espere", "depth": 3, "arguments": { "segundos": "3" } }, { "step": "salva_pagina", "depth": 3, "arguments": {} } ], "iterable": { "call": { "step": "localiza_elementos", "arguments": { "xpath": "'/html/body/div/div/div/div/div/div[2]/center/a[*]'", "numero_xpaths": "3" } } } } ], "iterable": { "call": { "step": "opcoes", "arguments": { "xpath": "'//*[@id=\"NUM_ANO_REFERENCIA\"]'" } } } } ] } ``` 2. Salve o coletor. 3. Edite esse coletor, retirando o parâmetro opcional `número de xpaths`. 4. Ao salvar essa edição, verificamos que esse parâmetro opcional que descartamos continua na configuração. ## Especificações da Coleta Exemplo: ``` { "source_name": "Licita\u00e7\u00f5es Itabirito", "base_url": "http://api.conectbr.com.br/Licitacao/Busca/?token=XEOk3i50lCa5IcpiNG7vXQ==", "obey_robots": false, "data_path": "/home/loui/coletas/licitacoes_itabirito", "request_type": "GET", "form_request_type": "POST", "antiblock_download_delay": 2, "antiblock_autothrottle_enabled": false, "antiblock_autothrottle_start_delay": 2, "antiblock_autothrottle_max_delay": 10, "antiblock_ip_rotation_enabled": false, "antiblock_ip_rotation_type": "tor", "antiblock_max_reqs_per_ip": 10, "antiblock_max_reuse_rounds": 10, "antiblock_proxy_list": "", "antiblock_user_agent_rotation_enabled": false, "antiblock_reqs_per_user_agent": 100, "antiblock_user_agents_list": "", "antiblock_insert_cookies_enabled": false, "antiblock_cookies_list": "", "captcha": "none", "has_webdriver": false, "webdriver_path": null, "img_xpath": null, "sound_xpath": null, "dynamic_processing": true, "explore_links": true, "link_extractor_max_depth": 1, "link_extractor_allow_url": "^(http:\\/\\/api.conectbr.com.br\\/Licitacao\\/AbreArquivo\\/.*)", "link_extractor_allow_domains": null, "link_extractor_tags": null, "link_extractor_attrs": null, "link_extractor_check_type": false, "link_extractor_process_value": "", "download_files": true, "download_files_allow_url": "^(http:\\/\\/api.conectbr.com.br\\/Licitacao\\/AbreArquivo\\/.*)", "download_files_allow_extensions": "pdf,doc", "download_files_allow_domains": null, "download_files_tags": null, "download_files_attrs": null, "download_files_process_value": "", "download_files_check_large_content": true, "download_imgs": false, "steps": "{\"step\":\"root\",\"depth\":0,\"children\":[{\"step\":\"espere\",\"depth\":1,\"arguments\":{\"segundos\":\"2\"}},{\"step\":\"para_cada\",\"depth\":1,\"iterator\":\"ano\",\"children\":[{\"step\":\"selecione\",\"depth\":2,\"arguments\":{\"xpath\":\"'//*[@id=\\\"NUM_ANO_REFERENCIA\\\"]'\",\"opcao\":\"ano\"}},{\"step\":\"clique\",\"depth\":2,\"arguments\":{\"elemento\":\"'//*[@id=\\\"formDados\\\"]/fieldset/div/div[7]/input'\"}},{\"step\":\"espere\",\"depth\":2,\"arguments\":{\"segundos\":\"3\"}},{\"step\":\"para_cada\",\"depth\":2,\"iterator\":\"pagina\",\"children\":[{\"step\":\"espere\",\"depth\":3,\"arguments\":{\"segundos\":\"3\"}},{\"step\":\"salva_pagina\",\"depth\":3,\"arguments\":{}}],\"iterable\":{\"call\":{\"step\":\"localiza_elementos\",\"arguments\":{\"xpath\":\"'/html/body/div/div/div/div/div/div[2]/center/a[*]'\",\"numero_xpaths\":\"3\"}}}}],\"iterable\":{\"call\":{\"step\":\"opcoes\",\"arguments\":{\"xpath\":\"'//*[@id=\\\"NUM_ANO_REFERENCIA\\\"]'\"}}}}]}", "encoding_detection_method": 1, "templated_url_parameter_handlers": [], "static_form_parameter_handlers": [], "templated_url_response_handlers": [], "static_form_response_handlers": [], "crawler_id": 3, "instance_id": "164202556313372" } ``` ## Sistema (caso necessário) - Problema encontrado na branch `master`, provavelmente presente no sistema distribuído e na versão das máquinas do MP.
process
problemas ao deletar parâmetros opcionais comportamento esperado espera se que parâmetros opcionais instanciados possam ser removidos quando desejamos retirá lo ao editarmos um coletor comportamento atual quando editamos um coletor descartando um parâmetro opcional de um passo dinâmico qualquer o sistema não salva essa alteração o parâmetro continua na configuração dos passos passos para reproduzir o erro crie um coletor utilizando algum passo que contenha um parâmetro opcional e instancie o um exemplo de um coletor com essas características está descrito na seção seguinte abaixo o json de passos referente a ele note que há o uso do parâmetro opcional número de xpaths do passo localizar elementos step root depth children step espere depth arguments segundos step para cada depth iterator ano children step selecione depth arguments xpath opcao ano step clique depth arguments elemento fieldset div div input step espere depth arguments segundos step para cada depth iterator pagina children step espere depth arguments segundos step salva pagina depth arguments iterable call step localiza elementos arguments xpath html body div div div div div div center a numero xpaths iterable call step opcoes arguments xpath salve o coletor edite esse coletor retirando o parâmetro opcional número de xpaths ao salvar essa edição verificamos que esse parâmetro opcional que descartamos continua na configuração especificações da coleta exemplo source name licita itabirito base url obey robots false data path home loui coletas licitacoes itabirito request type get form request type post antiblock download delay antiblock autothrottle enabled false antiblock autothrottle start delay antiblock autothrottle max delay antiblock ip rotation enabled false antiblock ip rotation type tor antiblock max reqs per ip antiblock max reuse rounds antiblock proxy list antiblock user agent rotation enabled false antiblock reqs per user agent antiblock user agents list antiblock insert cookies enabled false antiblock cookies list captcha none has webdriver false webdriver path null img xpath null sound xpath null dynamic processing true explore links true link extractor max depth link extractor allow url http api conectbr com br licitacao abrearquivo link extractor allow domains null link extractor tags null link extractor attrs null link extractor check type false link extractor process value download files true download files allow url http api conectbr com br licitacao abrearquivo download files allow extensions pdf doc download files allow domains null download files tags null download files attrs null download files process value download files check large content true download imgs false steps step root depth children opcao ano step clique depth arguments elemento fieldset div div input step espere depth arguments segundos step para cada depth iterator pagina children iterable call step localiza elementos arguments xpath html body div div div div div div center a numero xpaths iterable call step opcoes arguments xpath encoding detection method templated url parameter handlers static form parameter handlers templated url response handlers static form response handlers crawler id instance id sistema caso necessário problema encontrado na branch master provavelmente presente no sistema distribuído e na versão das máquinas do mp
1
17,993
24,011,188,895
IssuesEvent
2022-09-14 18:58:33
AcademySoftwareFoundation/OpenCue
https://api.github.com/repos/AcademySoftwareFoundation/OpenCue
opened
[cuegui] Upgrade to PySide6
process
**Describe the process** PySide2 is deprecated and no longer publishes wheels for newer platforms. I'm unable to `pip install PySide2` on my M1 macbook for example. This is blocking me from working on various other issues. We should upgrade to PySide6, which is the currently supported version.
1.0
[cuegui] Upgrade to PySide6 - **Describe the process** PySide2 is deprecated and no longer publishes wheels for newer platforms. I'm unable to `pip install PySide2` on my M1 macbook for example. This is blocking me from working on various other issues. We should upgrade to PySide6, which is the currently supported version.
process
upgrade to describe the process is deprecated and no longer publishes wheels for newer platforms i m unable to pip install on my macbook for example this is blocking me from working on various other issues we should upgrade to which is the currently supported version
1
171,231
27,086,759,020
IssuesEvent
2023-02-14 17:33:07
MetaMask/metamask-extension
https://api.github.com/repos/MetaMask/metamask-extension
closed
Update Story Links in Storybook Documentation
design-system team-extension-client team-extension-UX
When https://github.com/MetaMask/metamask-extension/pull/17518 was merged it changed some storybook URLs since the story name was changed and as a result, the link references to those stories are broken. We will need to update internal and external links. This ticket is to update all internal links inside of the repo. - A good start would be to search for `[(/docs` in `.mdx` files that point to component doc pages - Then search for `](/story` in `.mdx` files that point to component story pages - I would also do a very broad search `](/` and scan for any links that point to storybook pages ### Technical Details - Update all URLs to stories in `.mdx` files to use the new URLS ### Acceptance Criteria - All links to stories in MDX docs work
1.0
Update Story Links in Storybook Documentation - When https://github.com/MetaMask/metamask-extension/pull/17518 was merged it changed some storybook URLs since the story name was changed and as a result, the link references to those stories are broken. We will need to update internal and external links. This ticket is to update all internal links inside of the repo. - A good start would be to search for `[(/docs` in `.mdx` files that point to component doc pages - Then search for `](/story` in `.mdx` files that point to component story pages - I would also do a very broad search `](/` and scan for any links that point to storybook pages ### Technical Details - Update all URLs to stories in `.mdx` files to use the new URLS ### Acceptance Criteria - All links to stories in MDX docs work
non_process
update story links in storybook documentation when was merged it changed some storybook urls since the story name was changed and as a result the link references to those stories are broken we will need to update internal and external links this ticket is to update all internal links inside of the repo a good start would be to search for docs in mdx files that point to component doc pages then search for story in mdx files that point to component story pages i would also do a very broad search and scan for any links that point to storybook pages technical details update all urls to stories in mdx files to use the new urls acceptance criteria all links to stories in mdx docs work
0
257,111
22,147,084,466
IssuesEvent
2022-06-03 13:10:45
python/cpython
https://api.github.com/repos/python/cpython
closed
Attempt to further increase test coverage of calendar module
type-feature tests
BPO | [25528](https://bugs.python.org/issue25528) --- | :--- Nosy | @rhettinger, @vadmium, @serhiy-storchaka, @matrixise, @iritkatriel Files | <li>[mywork.patch](https://bugs.python.org/file40916/mywork.patch "Uploaded as text/plain at 2015-11-01.09:57:43 by Rohit Mediratta"): Patch to increase coverage for 3 additional lines</li><li>[mywork_update.patch](https://bugs.python.org/file41109/mywork_update.patch "Uploaded as text/plain at 2015-11-21.08:59:10 by Rohit Mediratta"): Updated patch showing the correct diff</li> <sup>*Note: these values reflect the state of the issue at the time it was migrated and might not reflect the current state.*</sup> <details><summary>Show more details</summary><p> GitHub fields: ```python assignee = None closed_at = None created_at = <Date 2015-11-01.09:57:43.791> labels = ['type-feature', 'tests'] title = 'Attempt to further increase test coverage of calendar module' updated_at = <Date 2022-03-22.21:27:14.499> user = 'https://bugs.python.org/RohitMediratta' ``` bugs.python.org fields: ```python activity = <Date 2022-03-22.21:27:14.499> actor = 'iritkatriel' assignee = 'none' closed = False closed_date = None closer = None components = ['Tests'] creation = <Date 2015-11-01.09:57:43.791> creator = 'Rohit Mediratta' dependencies = [] files = ['40916', '41109'] hgrepos = [] issue_num = 25528 keywords = ['patch'] message_count = 7.0 messages = ['253836', '254421', '255047', '255416', '255417', '255423', '415815'] nosy_count = 6.0 nosy_names = ['rhettinger', 'martin.panter', 'serhiy.storchaka', 'matrixise', 'Rohit Mediratta', 'iritkatriel'] pr_nums = [] priority = 'normal' resolution = None stage = 'patch review' status = 'open' superseder = None type = 'enhancement' url = 'https://bugs.python.org/issue25528' versions = ['Python 3.6'] ``` </p></details>
1.0
Attempt to further increase test coverage of calendar module - BPO | [25528](https://bugs.python.org/issue25528) --- | :--- Nosy | @rhettinger, @vadmium, @serhiy-storchaka, @matrixise, @iritkatriel Files | <li>[mywork.patch](https://bugs.python.org/file40916/mywork.patch "Uploaded as text/plain at 2015-11-01.09:57:43 by Rohit Mediratta"): Patch to increase coverage for 3 additional lines</li><li>[mywork_update.patch](https://bugs.python.org/file41109/mywork_update.patch "Uploaded as text/plain at 2015-11-21.08:59:10 by Rohit Mediratta"): Updated patch showing the correct diff</li> <sup>*Note: these values reflect the state of the issue at the time it was migrated and might not reflect the current state.*</sup> <details><summary>Show more details</summary><p> GitHub fields: ```python assignee = None closed_at = None created_at = <Date 2015-11-01.09:57:43.791> labels = ['type-feature', 'tests'] title = 'Attempt to further increase test coverage of calendar module' updated_at = <Date 2022-03-22.21:27:14.499> user = 'https://bugs.python.org/RohitMediratta' ``` bugs.python.org fields: ```python activity = <Date 2022-03-22.21:27:14.499> actor = 'iritkatriel' assignee = 'none' closed = False closed_date = None closer = None components = ['Tests'] creation = <Date 2015-11-01.09:57:43.791> creator = 'Rohit Mediratta' dependencies = [] files = ['40916', '41109'] hgrepos = [] issue_num = 25528 keywords = ['patch'] message_count = 7.0 messages = ['253836', '254421', '255047', '255416', '255417', '255423', '415815'] nosy_count = 6.0 nosy_names = ['rhettinger', 'martin.panter', 'serhiy.storchaka', 'matrixise', 'Rohit Mediratta', 'iritkatriel'] pr_nums = [] priority = 'normal' resolution = None stage = 'patch review' status = 'open' superseder = None type = 'enhancement' url = 'https://bugs.python.org/issue25528' versions = ['Python 3.6'] ``` </p></details>
non_process
attempt to further increase test coverage of calendar module bpo nosy rhettinger vadmium serhiy storchaka matrixise iritkatriel files uploaded as text plain at by rohit mediratta patch to increase coverage for additional lines uploaded as text plain at by rohit mediratta updated patch showing the correct diff note these values reflect the state of the issue at the time it was migrated and might not reflect the current state show more details github fields python assignee none closed at none created at labels title attempt to further increase test coverage of calendar module updated at user bugs python org fields python activity actor iritkatriel assignee none closed false closed date none closer none components creation creator rohit mediratta dependencies files hgrepos issue num keywords message count messages nosy count nosy names pr nums priority normal resolution none stage patch review status open superseder none type enhancement url versions
0
340,436
24,654,540,081
IssuesEvent
2022-10-17 21:48:52
kubernetes/minikube
https://api.github.com/repos/kubernetes/minikube
closed
document `socket_vmnet` installation
kind/documentation
### What Happened? We should document how to install `socket_vmnet` and point to this documentation when the user tries to use Qemu with `socket_vmnet` but doesn't actually have it installed.
1.0
document `socket_vmnet` installation - ### What Happened? We should document how to install `socket_vmnet` and point to this documentation when the user tries to use Qemu with `socket_vmnet` but doesn't actually have it installed.
non_process
document socket vmnet installation what happened we should document how to install socket vmnet and point to this documentation when the user tries to use qemu with socket vmnet but doesn t actually have it installed
0
329,217
28,208,620,370
IssuesEvent
2023-04-05 00:48:42
golang/go
https://api.github.com/repos/golang/go
closed
cmd/dist: linker warnings when compiling misc/cgo tests statically on `boringcrypto` builder
Testing NeedsFix
``` #!watchflakes post <- log ~ `warning: Using 'getaddrinfo' in statically linked applications` ``` Issue created automatically to collect these failures. Example ([log](https://build.golang.org/log/dc7a10dbecbaecf752ec9d3ab8ce7d6bc7707be0)): /usr/bin/ld: /workdir/tmp/go-link-1495653376/000023.o: in function `bio_ip_and_port_to_socket_and_addr': (.text+0xcc2f9): warning: Using 'getaddrinfo' in statically linked applications requires at runtime the shared libraries from the glibc version used for linking — [watchflakes](https://go.dev/wiki/Watchflakes)
1.0
cmd/dist: linker warnings when compiling misc/cgo tests statically on `boringcrypto` builder - ``` #!watchflakes post <- log ~ `warning: Using 'getaddrinfo' in statically linked applications` ``` Issue created automatically to collect these failures. Example ([log](https://build.golang.org/log/dc7a10dbecbaecf752ec9d3ab8ce7d6bc7707be0)): /usr/bin/ld: /workdir/tmp/go-link-1495653376/000023.o: in function `bio_ip_and_port_to_socket_and_addr': (.text+0xcc2f9): warning: Using 'getaddrinfo' in statically linked applications requires at runtime the shared libraries from the glibc version used for linking — [watchflakes](https://go.dev/wiki/Watchflakes)
non_process
cmd dist linker warnings when compiling misc cgo tests statically on boringcrypto builder watchflakes post log warning using getaddrinfo in statically linked applications issue created automatically to collect these failures example usr bin ld workdir tmp go link o in function bio ip and port to socket and addr text warning using getaddrinfo in statically linked applications requires at runtime the shared libraries from the glibc version used for linking —
0
66,477
14,785,019,767
IssuesEvent
2021-01-12 01:40:32
Watemlifts/pritunl
https://api.github.com/repos/Watemlifts/pritunl
opened
CVE-2020-15366 (Medium) detected in ajv-5.5.2.tgz
security vulnerability
## CVE-2020-15366 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ajv-5.5.2.tgz</b></p></summary> <p>Another JSON Schema Validator</p> <p>Library home page: <a href="https://registry.npmjs.org/ajv/-/ajv-5.5.2.tgz">https://registry.npmjs.org/ajv/-/ajv-5.5.2.tgz</a></p> <p>Path to dependency file: pritunl/www/package.json</p> <p>Path to vulnerable library: pritunl/www/node_modules/ajv/package.json</p> <p> Dependency Hierarchy: - grunt-contrib-jshint-1.1.0.tgz (Root Library) - jshint-2.9.6.tgz - phantomjs-prebuilt-2.1.16.tgz - request-2.87.0.tgz - har-validator-5.0.3.tgz - :x: **ajv-5.5.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Watemlifts/pritunl/commit/f33d93230e92dc0082d2e75ce28b0ee61d409141">f33d93230e92dc0082d2e75ce28b0ee61d409141</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in ajv.validate() in Ajv (aka Another JSON Schema Validator) 6.12.2. A carefully crafted JSON schema could be provided that allows execution of other code by prototype pollution. (While untrusted schemas are recommended against, the worst case of an untrusted schema should be a denial of service, not execution of code.) <p>Publish Date: 2020-07-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15366>CVE-2020-15366</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/ajv-validator/ajv/releases/tag/v6.12.3">https://github.com/ajv-validator/ajv/releases/tag/v6.12.3</a></p> <p>Release Date: 2020-07-15</p> <p>Fix Resolution: ajv - 6.12.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-15366 (Medium) detected in ajv-5.5.2.tgz - ## CVE-2020-15366 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ajv-5.5.2.tgz</b></p></summary> <p>Another JSON Schema Validator</p> <p>Library home page: <a href="https://registry.npmjs.org/ajv/-/ajv-5.5.2.tgz">https://registry.npmjs.org/ajv/-/ajv-5.5.2.tgz</a></p> <p>Path to dependency file: pritunl/www/package.json</p> <p>Path to vulnerable library: pritunl/www/node_modules/ajv/package.json</p> <p> Dependency Hierarchy: - grunt-contrib-jshint-1.1.0.tgz (Root Library) - jshint-2.9.6.tgz - phantomjs-prebuilt-2.1.16.tgz - request-2.87.0.tgz - har-validator-5.0.3.tgz - :x: **ajv-5.5.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Watemlifts/pritunl/commit/f33d93230e92dc0082d2e75ce28b0ee61d409141">f33d93230e92dc0082d2e75ce28b0ee61d409141</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in ajv.validate() in Ajv (aka Another JSON Schema Validator) 6.12.2. A carefully crafted JSON schema could be provided that allows execution of other code by prototype pollution. (While untrusted schemas are recommended against, the worst case of an untrusted schema should be a denial of service, not execution of code.) <p>Publish Date: 2020-07-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15366>CVE-2020-15366</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/ajv-validator/ajv/releases/tag/v6.12.3">https://github.com/ajv-validator/ajv/releases/tag/v6.12.3</a></p> <p>Release Date: 2020-07-15</p> <p>Fix Resolution: ajv - 6.12.3</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in ajv tgz cve medium severity vulnerability vulnerable library ajv tgz another json schema validator library home page a href path to dependency file pritunl www package json path to vulnerable library pritunl www node modules ajv package json dependency hierarchy grunt contrib jshint tgz root library jshint tgz phantomjs prebuilt tgz request tgz har validator tgz x ajv tgz vulnerable library found in head commit a href vulnerability details an issue was discovered in ajv validate in ajv aka another json schema validator a carefully crafted json schema could be provided that allows execution of other code by prototype pollution while untrusted schemas are recommended against the worst case of an untrusted schema should be a denial of service not execution of code publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ajv step up your open source security game with whitesource
0
7,392
6,022,516,221
IssuesEvent
2017-06-07 21:14:50
PowerShell/PowerShell
https://api.github.com/repos/PowerShell/PowerShell
closed
Powershell initialization on linux is slow
Area-Performance Resolution-Fixed
## Steps to reproduce Launch powershell ## Expected behavior Usable prompt in milliseconds ## Actual behavior Usable prompt in seconds ## Environment data ``` powershell PS /> $PSVersionTable Name Value ---- ----- PSVersion 6.0.0-alpha PSEdition Core PSCompatibleVersions {1.0, 2.0, 3.0, 4.0...} BuildVersion 3.0.0.0 GitCommitId v6.0.0-alpha.9 CLRVersion WSManStackVersion 3.0 PSRemotingProtocolVersion 2.3 SerializationVersion 1.1.0.1 ``` For example: ``` [me@ipsum] ~ ➜ cat test.ps1 #!/usr/bin/powershell exit [me@ipsum] ~ ➜ time ./test.ps1 ./test.ps1 1.30s user 0.12s system 138% cpu 1.035 total ``` Compared to: ``` [me@ipsum] ~ ➜ cat test.sh #/bin/bash exit [me@ipsum] ~ ➜ time ./test.sh ./test.sh 0.00s user 0.00s system 0% cpu 0.002 total ``` The windows host I have available: ``` PS C:\Users\me> echo "exit" > test.ps1 PS C:\Users\me> Measure-Command {.\test.ps1} Days : 0 Hours : 0 Minutes : 0 Seconds : 0 Milliseconds : 7 Ticks : 71141 TotalDays : 8.23391203703704E-08 TotalHours : 1.97613888888889E-06 TotalMinutes : 0.000118568333333333 TotalSeconds : 0.0071141 TotalMilliseconds : 7.1141 PS C:\Users\me> $PSVersionTable Name Value ---- ----- PSVersion 3.0 WSManStackVersion 3.0 SerializationVersion 1.1.0.1 CLRVersion 4.0.30319.18033 BuildVersion 6.2.9200.16434 PSCompatibleVersions {1.0, 2.0, 3.0} PSRemotingProtocolVersion 2.2 ```
True
Powershell initialization on linux is slow - ## Steps to reproduce Launch powershell ## Expected behavior Usable prompt in milliseconds ## Actual behavior Usable prompt in seconds ## Environment data ``` powershell PS /> $PSVersionTable Name Value ---- ----- PSVersion 6.0.0-alpha PSEdition Core PSCompatibleVersions {1.0, 2.0, 3.0, 4.0...} BuildVersion 3.0.0.0 GitCommitId v6.0.0-alpha.9 CLRVersion WSManStackVersion 3.0 PSRemotingProtocolVersion 2.3 SerializationVersion 1.1.0.1 ``` For example: ``` [me@ipsum] ~ ➜ cat test.ps1 #!/usr/bin/powershell exit [me@ipsum] ~ ➜ time ./test.ps1 ./test.ps1 1.30s user 0.12s system 138% cpu 1.035 total ``` Compared to: ``` [me@ipsum] ~ ➜ cat test.sh #/bin/bash exit [me@ipsum] ~ ➜ time ./test.sh ./test.sh 0.00s user 0.00s system 0% cpu 0.002 total ``` The windows host I have available: ``` PS C:\Users\me> echo "exit" > test.ps1 PS C:\Users\me> Measure-Command {.\test.ps1} Days : 0 Hours : 0 Minutes : 0 Seconds : 0 Milliseconds : 7 Ticks : 71141 TotalDays : 8.23391203703704E-08 TotalHours : 1.97613888888889E-06 TotalMinutes : 0.000118568333333333 TotalSeconds : 0.0071141 TotalMilliseconds : 7.1141 PS C:\Users\me> $PSVersionTable Name Value ---- ----- PSVersion 3.0 WSManStackVersion 3.0 SerializationVersion 1.1.0.1 CLRVersion 4.0.30319.18033 BuildVersion 6.2.9200.16434 PSCompatibleVersions {1.0, 2.0, 3.0} PSRemotingProtocolVersion 2.2 ```
non_process
powershell initialization on linux is slow steps to reproduce launch powershell expected behavior usable prompt in milliseconds actual behavior usable prompt in seconds environment data powershell ps psversiontable name value psversion alpha psedition core pscompatibleversions buildversion gitcommitid alpha clrversion wsmanstackversion psremotingprotocolversion serializationversion for example ➜ cat test usr bin powershell exit ➜ time test test user system cpu total compared to ➜ cat test sh bin bash exit ➜ time test sh test sh user system cpu total the windows host i have available ps c users me echo exit test ps c users me measure command test days hours minutes seconds milliseconds ticks totaldays totalhours totalminutes totalseconds totalmilliseconds ps c users me psversiontable name value psversion wsmanstackversion serializationversion clrversion buildversion pscompatibleversions psremotingprotocolversion
0
11,067
13,903,469,866
IssuesEvent
2020-10-20 07:17:43
decidim/decidim
https://api.github.com/repos/decidim/decidim
closed
Analyze ContentBlocks for adapting them to Process Groups
contract: process-groups
Analyze and implement changes if needed to Blocks internals to allow usage in Process Groups context.
1.0
Analyze ContentBlocks for adapting them to Process Groups - Analyze and implement changes if needed to Blocks internals to allow usage in Process Groups context.
process
analyze contentblocks for adapting them to process groups analyze and implement changes if needed to blocks internals to allow usage in process groups context
1
65,263
16,175,041,797
IssuesEvent
2021-05-03 04:34:45
hash-checker/hash-checker-lite
https://api.github.com/repos/hash-checker/hash-checker-lite
closed
An error occurred during gradle syncing
build error
![gradle_issue](https://user-images.githubusercontent.com/32436953/116808727-1fd2cf80-ab43-11eb-8dc1-904d3884254e.png) Changing System.getenv(KEYSTORE_PATH) to String.valueOf(System.getenv(KEYSTORE_PATH)) solves the problem and makes a successful gradle sync The same for the other three calls of System.getenv If you accept this, assign it to me, please
1.0
An error occurred during gradle syncing - ![gradle_issue](https://user-images.githubusercontent.com/32436953/116808727-1fd2cf80-ab43-11eb-8dc1-904d3884254e.png) Changing System.getenv(KEYSTORE_PATH) to String.valueOf(System.getenv(KEYSTORE_PATH)) solves the problem and makes a successful gradle sync The same for the other three calls of System.getenv If you accept this, assign it to me, please
non_process
an error occurred during gradle syncing changing system getenv keystore path to string valueof system getenv keystore path solves the problem and makes a successful gradle sync the same for the other three calls of system getenv if you accept this assign it to me please
0
19,679
26,031,908,329
IssuesEvent
2022-12-21 22:20:22
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
Description of $(Rev:r) lacks meaning of $(Rev:.r) and its used on this page in most of examples
doc-enhancement devops/prod Pri2 devops-cicd-process/tech
In $(Rev:r) definition there is no mention of adding dot, just r or rr as parameter. Whats the meaning of $(Rev:.r) and how does it differs from $(Rev:r) Thanks --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: a57f8545-bb15-3a71-1876-3a9ec1a59b93 * Version Independent ID: 28c87c8d-c28d-7493-0c7c-8c38b04fbcd7 * Content: [Run (build) number - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/run-number?view=azure-devops&tabs=yaml&source=docs) * Content Source: [docs/pipelines/process/run-number.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/run-number.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
Description of $(Rev:r) lacks meaning of $(Rev:.r) and its used on this page in most of examples - In $(Rev:r) definition there is no mention of adding dot, just r or rr as parameter. Whats the meaning of $(Rev:.r) and how does it differs from $(Rev:r) Thanks --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: a57f8545-bb15-3a71-1876-3a9ec1a59b93 * Version Independent ID: 28c87c8d-c28d-7493-0c7c-8c38b04fbcd7 * Content: [Run (build) number - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/run-number?view=azure-devops&tabs=yaml&source=docs) * Content Source: [docs/pipelines/process/run-number.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/run-number.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
description of rev r lacks meaning of rev r and its used on this page in most of examples in rev r definition there is no mention of adding dot just r or rr as parameter whats the meaning of rev r and how does it differs from rev r thanks document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
18,635
24,580,448,350
IssuesEvent
2022-10-13 15:14:31
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[iOS] [FHIR] Questionnaire response > Other answers are getting stored for the form step questions in the following scenario
Bug P1 iOS Response datastore Process: Fixed Process: Tested QA Process: Tested dev
Steps: 1. In SB add multiple text choice questions in the form of step 2. In mobile, provide the response only for the first question 3. Don't provide the answer to the 2nd question 4. Submit the response 5. Go to the FHIR store and observe AR: Selected value for the first question in the form is getting displayed for the skipped question ER: Answer options should get displayed properly ![image](https://user-images.githubusercontent.com/71445210/192263421-f7d2ea22-b69e-4681-b1b7-fcdb9fdf3f06.png)
3.0
[iOS] [FHIR] Questionnaire response > Other answers are getting stored for the form step questions in the following scenario - Steps: 1. In SB add multiple text choice questions in the form of step 2. In mobile, provide the response only for the first question 3. Don't provide the answer to the 2nd question 4. Submit the response 5. Go to the FHIR store and observe AR: Selected value for the first question in the form is getting displayed for the skipped question ER: Answer options should get displayed properly ![image](https://user-images.githubusercontent.com/71445210/192263421-f7d2ea22-b69e-4681-b1b7-fcdb9fdf3f06.png)
process
questionnaire response other answers are getting stored for the form step questions in the following scenario steps in sb add multiple text choice questions in the form of step in mobile provide the response only for the first question don t provide the answer to the question submit the response go to the fhir store and observe ar selected value for the first question in the form is getting displayed for the skipped question er answer options should get displayed properly
1
430,840
30,203,166,154
IssuesEvent
2023-07-05 07:38:09
d702e20/CGAAL
https://api.github.com/repos/d702e20/CGAAL
closed
The JSON example models/queries are wrong
documentation good first issue
For instance, the model `json-examples/mexican_standoff/mexican_standoff.json` with query `json-examples/mexican_standoff/can_3_players_gurantee_atleast_one_of_them_survives.json` results in false, but should be true. Looks like labels might be off-by-one. But there are more problems: The query `json-examples/mexican_standoff/can_3_players_gurantee_atleast_one_of_them_survives.atl` which contains `<<p1,p2,p3>> G (p1.alive | p2.alive | p3.alive)` which cannot even parse, since json models do not have a concept of `p1` or `alive`. These should be fixed.
1.0
The JSON example models/queries are wrong - For instance, the model `json-examples/mexican_standoff/mexican_standoff.json` with query `json-examples/mexican_standoff/can_3_players_gurantee_atleast_one_of_them_survives.json` results in false, but should be true. Looks like labels might be off-by-one. But there are more problems: The query `json-examples/mexican_standoff/can_3_players_gurantee_atleast_one_of_them_survives.atl` which contains `<<p1,p2,p3>> G (p1.alive | p2.alive | p3.alive)` which cannot even parse, since json models do not have a concept of `p1` or `alive`. These should be fixed.
non_process
the json example models queries are wrong for instance the model json examples mexican standoff mexican standoff json with query json examples mexican standoff can players gurantee atleast one of them survives json results in false but should be true looks like labels might be off by one but there are more problems the query json examples mexican standoff can players gurantee atleast one of them survives atl which contains g alive alive alive which cannot even parse since json models do not have a concept of or alive these should be fixed
0
112,780
24,327,746,207
IssuesEvent
2022-09-30 16:15:43
sourcegraph/sourcegraph
https://api.github.com/repos/sourcegraph/sourcegraph
opened
insights: ensure pings are emitted successfully
team/code-insights backend
This an open ended problem statement, but as a follow up to our need to patch 4.01 with ping fixes. Consider the following (or other ideas): 1. alert on ping objects 2. automated tests for each ping object
1.0
insights: ensure pings are emitted successfully - This an open ended problem statement, but as a follow up to our need to patch 4.01 with ping fixes. Consider the following (or other ideas): 1. alert on ping objects 2. automated tests for each ping object
non_process
insights ensure pings are emitted successfully this an open ended problem statement but as a follow up to our need to patch with ping fixes consider the following or other ideas alert on ping objects automated tests for each ping object
0
10,426
13,218,181,134
IssuesEvent
2020-08-17 08:16:21
bisq-network/proposals
https://api.github.com/repos/bisq-network/proposals
closed
[WIP] User support
an:idea re:processes was:superseded
> _This is a Bisq Network proposal. Please familiarize yourself with the [submission and review process](https://docs.bisq.network/proposals.html)._ <!-- Please do not remove the text above. --> **Problem** When a user creates a support issue, the developer providing support has 2 goals in mind: 1. Help the user solve her problem. 2. Prevent other users from facing the same problem. Sometimes both goals are achieved by doing just one task, e.g. solving the bug. Sometimes there is a quick workaround for the user but the bug is difficult to reproduce/fix. Then, there are some other challenges: - Developers tend to help user use a workaround (e.g. do a restore from seed for a corrupted wallet) and skip finding the bug that caused the problem. Developers are reluctant to try to find the bug that caused user problem because many times it is a dead end or it can take a huge effort to reproduce the bug. - Eg https://github.com/bisq-network/bisq/issues/2778 - User bug report is generally not exhaustive: they usually write something like “I have a problem buying bitcoin” while the developer expects a step-by-step list of what they did and what they got, the logs, the bisq version she is using, etc. - The user is just interested in getting her problem solved and sometimes are not very eager to cooperate to fix the bug. - On most cases, the user takes hours/days to reply and the developer can’t talk to the user over the phone. - The user usually reports the problem after trying several times, that makes it difficult to find the original cause. - There are plenty of things logged as ERROR on the logs. Most of them are not real errors (mostly connection problems), which makes it hard to discover the real errors. - Full stack trace not always present in the logs. - Sometimes the full stack trace is not present on the logs because code like this is used: - `e.printStackTrace()` - prints to the console, not the log file. - `throw new CryptoException("Signature verification failed" + e.getMessage())` - just prints the original message, not the stack trace. - Some developers are specialists in some part of bisq and don't know how the rest of the app works so it is difficult to solve bugs that are outside their domain. - User confidential information. The developer might need the latest few lines of the user log, the entire user log and/or the user's bisq data dir to reproduce the bug. Sharing that info means to the user exposing her privacy. **Proposals** - Create issue templates for user bugs - similar to the ones used in the support github project (See https://github.com/bisq-network/support/issues/new?assignees=&labels=&template=reimbursement-template.md&title=Fee+reimbursement+for+trade+XXXXX). - The developer should not provide the user a workaround until the user has provided the developer all the information needed to reproduce the bug. - Improve the log - Replace `log.error()` by `log.warn()` for connectivity problems that are not bugs. - Replace `e.printStackTrace()` by `log.warn("Some text", e)`. - Replace `throw new SomeException("Some text" + e.getMessage())` by `throw new SomeException("Some text", e)` - Design a procedure for devs to get help from other devs when solving issues. - Design a procedure for user sending confidential information. Here is an idea: User uploads to dropbox (or any other app) the last few line of the logs, the entire logs and/or the bisq data dir and sends an email to the developer working on the case with a link to the confidential information. User should be able to distinguish a real developer vs a fake one just trying to get user’s private data. User is still exposing private info to the developer. Developer should help the user withdraw her btc balance before doing that. Problem: BSQ balance can not be withdrawn. - Have support layers. - Level 1 - Helps the user if the problem is not a bug but the user does not know how to use bisq. - Otherwise, helps the user to complete the issue in a more thorough way and assigns the issue to the relevant developer. - Level 2 - Developers with experience in some bisq areas investigate and fix the bug. - Design clear compensation incentives for different support tasks. - Providing a workaround to the user. - Investigating the issue but not being able to reproduce the bug. - Investigating the issue and being able to reproduce the bug and document it. - Fixing the bug.
1.0
[WIP] User support - > _This is a Bisq Network proposal. Please familiarize yourself with the [submission and review process](https://docs.bisq.network/proposals.html)._ <!-- Please do not remove the text above. --> **Problem** When a user creates a support issue, the developer providing support has 2 goals in mind: 1. Help the user solve her problem. 2. Prevent other users from facing the same problem. Sometimes both goals are achieved by doing just one task, e.g. solving the bug. Sometimes there is a quick workaround for the user but the bug is difficult to reproduce/fix. Then, there are some other challenges: - Developers tend to help user use a workaround (e.g. do a restore from seed for a corrupted wallet) and skip finding the bug that caused the problem. Developers are reluctant to try to find the bug that caused user problem because many times it is a dead end or it can take a huge effort to reproduce the bug. - Eg https://github.com/bisq-network/bisq/issues/2778 - User bug report is generally not exhaustive: they usually write something like “I have a problem buying bitcoin” while the developer expects a step-by-step list of what they did and what they got, the logs, the bisq version she is using, etc. - The user is just interested in getting her problem solved and sometimes are not very eager to cooperate to fix the bug. - On most cases, the user takes hours/days to reply and the developer can’t talk to the user over the phone. - The user usually reports the problem after trying several times, that makes it difficult to find the original cause. - There are plenty of things logged as ERROR on the logs. Most of them are not real errors (mostly connection problems), which makes it hard to discover the real errors. - Full stack trace not always present in the logs. - Sometimes the full stack trace is not present on the logs because code like this is used: - `e.printStackTrace()` - prints to the console, not the log file. - `throw new CryptoException("Signature verification failed" + e.getMessage())` - just prints the original message, not the stack trace. - Some developers are specialists in some part of bisq and don't know how the rest of the app works so it is difficult to solve bugs that are outside their domain. - User confidential information. The developer might need the latest few lines of the user log, the entire user log and/or the user's bisq data dir to reproduce the bug. Sharing that info means to the user exposing her privacy. **Proposals** - Create issue templates for user bugs - similar to the ones used in the support github project (See https://github.com/bisq-network/support/issues/new?assignees=&labels=&template=reimbursement-template.md&title=Fee+reimbursement+for+trade+XXXXX). - The developer should not provide the user a workaround until the user has provided the developer all the information needed to reproduce the bug. - Improve the log - Replace `log.error()` by `log.warn()` for connectivity problems that are not bugs. - Replace `e.printStackTrace()` by `log.warn("Some text", e)`. - Replace `throw new SomeException("Some text" + e.getMessage())` by `throw new SomeException("Some text", e)` - Design a procedure for devs to get help from other devs when solving issues. - Design a procedure for user sending confidential information. Here is an idea: User uploads to dropbox (or any other app) the last few line of the logs, the entire logs and/or the bisq data dir and sends an email to the developer working on the case with a link to the confidential information. User should be able to distinguish a real developer vs a fake one just trying to get user’s private data. User is still exposing private info to the developer. Developer should help the user withdraw her btc balance before doing that. Problem: BSQ balance can not be withdrawn. - Have support layers. - Level 1 - Helps the user if the problem is not a bug but the user does not know how to use bisq. - Otherwise, helps the user to complete the issue in a more thorough way and assigns the issue to the relevant developer. - Level 2 - Developers with experience in some bisq areas investigate and fix the bug. - Design clear compensation incentives for different support tasks. - Providing a workaround to the user. - Investigating the issue but not being able to reproduce the bug. - Investigating the issue and being able to reproduce the bug and document it. - Fixing the bug.
process
user support this is a bisq network proposal please familiarize yourself with the problem when a user creates a support issue the developer providing support has goals in mind help the user solve her problem prevent other users from facing the same problem sometimes both goals are achieved by doing just one task e g solving the bug sometimes there is a quick workaround for the user but the bug is difficult to reproduce fix then there are some other challenges developers tend to help user use a workaround e g do a restore from seed for a corrupted wallet and skip finding the bug that caused the problem developers are reluctant to try to find the bug that caused user problem because many times it is a dead end or it can take a huge effort to reproduce the bug eg user bug report is generally not exhaustive they usually write something like “i have a problem buying bitcoin” while the developer expects a step by step list of what they did and what they got the logs the bisq version she is using etc the user is just interested in getting her problem solved and sometimes are not very eager to cooperate to fix the bug on most cases the user takes hours days to reply and the developer can’t talk to the user over the phone the user usually reports the problem after trying several times that makes it difficult to find the original cause there are plenty of things logged as error on the logs most of them are not real errors mostly connection problems which makes it hard to discover the real errors full stack trace not always present in the logs sometimes the full stack trace is not present on the logs because code like this is used e printstacktrace prints to the console not the log file throw new cryptoexception signature verification failed e getmessage just prints the original message not the stack trace some developers are specialists in some part of bisq and don t know how the rest of the app works so it is difficult to solve bugs that are outside their domain user confidential information the developer might need the latest few lines of the user log the entire user log and or the user s bisq data dir to reproduce the bug sharing that info means to the user exposing her privacy proposals create issue templates for user bugs similar to the ones used in the support github project see the developer should not provide the user a workaround until the user has provided the developer all the information needed to reproduce the bug improve the log replace log error by log warn for connectivity problems that are not bugs replace e printstacktrace by log warn some text e replace throw new someexception some text e getmessage by throw new someexception some text e design a procedure for devs to get help from other devs when solving issues design a procedure for user sending confidential information here is an idea user uploads to dropbox or any other app the last few line of the logs the entire logs and or the bisq data dir and sends an email to the developer working on the case with a link to the confidential information user should be able to distinguish a real developer vs a fake one just trying to get user’s private data user is still exposing private info to the developer developer should help the user withdraw her btc balance before doing that problem bsq balance can not be withdrawn have support layers level helps the user if the problem is not a bug but the user does not know how to use bisq otherwise helps the user to complete the issue in a more thorough way and assigns the issue to the relevant developer level developers with experience in some bisq areas investigate and fix the bug design clear compensation incentives for different support tasks providing a workaround to the user investigating the issue but not being able to reproduce the bug investigating the issue and being able to reproduce the bug and document it fixing the bug
1
414,531
27,991,100,695
IssuesEvent
2023-03-27 03:53:15
CarsOk/castillo_del_rey
https://api.github.com/repos/CarsOk/castillo_del_rey
closed
pagina de diagrama
documentation
## Yo como Aprendiz Sena ##Quiero Realizar pagina de diagramas ##Para Seguimientos del proyecto
1.0
pagina de diagrama - ## Yo como Aprendiz Sena ##Quiero Realizar pagina de diagramas ##Para Seguimientos del proyecto
non_process
pagina de diagrama yo como aprendiz sena quiero realizar pagina de diagramas para seguimientos del proyecto
0
20,313
26,956,267,760
IssuesEvent
2023-02-08 15:06:52
carbon-design-system/ibm-cloud-cognitive
https://api.github.com/repos/carbon-design-system/ibm-cloud-cognitive
closed
Reduce frequency of update workflow
type: process improvement
## What will this achieve? With contributors’ time split between product and library, we should reduce dependency updates to be once a month for the time being. ## How will success be measured? ## Additional information - Designs - Existing code - etc
1.0
Reduce frequency of update workflow - ## What will this achieve? With contributors’ time split between product and library, we should reduce dependency updates to be once a month for the time being. ## How will success be measured? ## Additional information - Designs - Existing code - etc
process
reduce frequency of update workflow what will this achieve with contributors’ time split between product and library we should reduce dependency updates to be once a month for the time being how will success be measured additional information designs existing code etc
1
223,073
17,102,639,207
IssuesEvent
2021-07-09 13:28:00
learning-at-home/hivemind
https://api.github.com/repos/learning-at-home/hivemind
opened
Document relevant parts of libp2p
documentation
Currently, all interfaces with libp2p only gloss over the inner workings of this library, which might not be very helpful for our contributors that want to understand the design decisions behind hivemind that are connected to libp2p integration. For instance, we possibly want to document that: * Connections to different P2P multiaddrs are deduplicated if they correspond to the same running daemon * Each peer might have several multiaddrs (basically, different endpoints) * Something else? I suggest to add a page with relevant parts of libp2p (possibly with links to upstream documentation) and specifics of interaction between this library and hivemind.
1.0
Document relevant parts of libp2p - Currently, all interfaces with libp2p only gloss over the inner workings of this library, which might not be very helpful for our contributors that want to understand the design decisions behind hivemind that are connected to libp2p integration. For instance, we possibly want to document that: * Connections to different P2P multiaddrs are deduplicated if they correspond to the same running daemon * Each peer might have several multiaddrs (basically, different endpoints) * Something else? I suggest to add a page with relevant parts of libp2p (possibly with links to upstream documentation) and specifics of interaction between this library and hivemind.
non_process
document relevant parts of currently all interfaces with only gloss over the inner workings of this library which might not be very helpful for our contributors that want to understand the design decisions behind hivemind that are connected to integration for instance we possibly want to document that connections to different multiaddrs are deduplicated if they correspond to the same running daemon each peer might have several multiaddrs basically different endpoints something else i suggest to add a page with relevant parts of possibly with links to upstream documentation and specifics of interaction between this library and hivemind
0
14,630
17,767,716,567
IssuesEvent
2021-08-30 09:40:38
googleapis/nodejs-asset
https://api.github.com/repos/googleapis/nodejs-asset
closed
Dependency Dashboard
type: process api: cloudasset
This issue contains a list of Renovate updates and their statuses. ## Awaiting Schedule These updates are awaiting their schedule. Click on a checkbox to get an update now. - [ ] <!-- unschedule-branch=renovate/actions-setup-node-2.x -->chore(deps): update actions/setup-node action to v2 ## Ignored or Blocked These are blocked by an existing closed PR and will not be recreated unless you click a checkbox below. - [ ] <!-- recreate-branch=renovate/gts-3.x -->[chore(deps): update dependency gts to v3](../pull/403) - [ ] <!-- recreate-branch=renovate/mocha-9.x -->[chore(deps): update dependency mocha to v9](../pull/498) (`mocha`, `@types/mocha`) - [ ] <!-- recreate-branch=renovate/yargs-17.x -->[fix(deps): update dependency yargs to v17](../pull/480) --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
1.0
Dependency Dashboard - This issue contains a list of Renovate updates and their statuses. ## Awaiting Schedule These updates are awaiting their schedule. Click on a checkbox to get an update now. - [ ] <!-- unschedule-branch=renovate/actions-setup-node-2.x -->chore(deps): update actions/setup-node action to v2 ## Ignored or Blocked These are blocked by an existing closed PR and will not be recreated unless you click a checkbox below. - [ ] <!-- recreate-branch=renovate/gts-3.x -->[chore(deps): update dependency gts to v3](../pull/403) - [ ] <!-- recreate-branch=renovate/mocha-9.x -->[chore(deps): update dependency mocha to v9](../pull/498) (`mocha`, `@types/mocha`) - [ ] <!-- recreate-branch=renovate/yargs-17.x -->[fix(deps): update dependency yargs to v17](../pull/480) --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
process
dependency dashboard this issue contains a list of renovate updates and their statuses awaiting schedule these updates are awaiting their schedule click on a checkbox to get an update now chore deps update actions setup node action to ignored or blocked these are blocked by an existing closed pr and will not be recreated unless you click a checkbox below pull pull mocha types mocha pull check this box to trigger a request for renovate to run again on this repository
1
340,399
30,513,899,487
IssuesEvent
2023-07-19 00:07:05
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
cloud/amazon: TestReadFileAtReturnsSize failed
C-test-failure O-robot branch-master T-disaster-recovery
cloud/amazon.TestReadFileAtReturnsSize [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_CloudUnitTests/10966069?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_CloudUnitTests/10966069?buildTab=artifacts#/) on master @ [06b5ba7888339ed53477ff73f9f3ae77a708ee17](https://github.com/cockroachdb/cockroach/commits/06b5ba7888339ed53477ff73f9f3ae77a708ee17): ``` === RUN TestReadFileAtReturnsSize s3_storage_test.go:639: Error Trace: github.com/cockroachdb/cockroach/pkg/cloud/amazon/s3_storage_test.go:639 Error: Not equal: expected: 11 actual : 0 Test: TestReadFileAtReturnsSize --- FAIL: TestReadFileAtReturnsSize (0.04s) ``` <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> /cc @cockroachdb/disaster-recovery <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestReadFileAtReturnsSize.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
1.0
cloud/amazon: TestReadFileAtReturnsSize failed - cloud/amazon.TestReadFileAtReturnsSize [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_CloudUnitTests/10966069?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_CloudUnitTests/10966069?buildTab=artifacts#/) on master @ [06b5ba7888339ed53477ff73f9f3ae77a708ee17](https://github.com/cockroachdb/cockroach/commits/06b5ba7888339ed53477ff73f9f3ae77a708ee17): ``` === RUN TestReadFileAtReturnsSize s3_storage_test.go:639: Error Trace: github.com/cockroachdb/cockroach/pkg/cloud/amazon/s3_storage_test.go:639 Error: Not equal: expected: 11 actual : 0 Test: TestReadFileAtReturnsSize --- FAIL: TestReadFileAtReturnsSize (0.04s) ``` <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> /cc @cockroachdb/disaster-recovery <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestReadFileAtReturnsSize.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
non_process
cloud amazon testreadfileatreturnssize failed cloud amazon testreadfileatreturnssize with on master run testreadfileatreturnssize storage test go error trace github com cockroachdb cockroach pkg cloud amazon storage test go error not equal expected actual test testreadfileatreturnssize fail testreadfileatreturnssize help see also cc cockroachdb disaster recovery
0
10,680
13,463,472,503
IssuesEvent
2020-09-09 17:38:57
chavarera/python-mini-projects
https://api.github.com/repos/chavarera/python-mini-projects
closed
Convert JPEG file to PNG
Assigned Image-processing Learn good first issue intermediate
# Description Script to convert JPEG file into PNG format. ## Type of issue - [x] Feature (New Script) ## Checklist: - [x] I have read the project guidelines. - [x] I have checked all the existing projects, before submitting a new project issue. - [x] I have checked previous issues to avoid duplicates. - [x] This issue will be meaningful for the project. <!-- Uncomment this in case you have a issue related to a bug in existing code.--> <!-- - [ ] I have added screenshots of the bug - [ ] I have added steps to reproduce the bug - [ ] I have proposed a possible solution for the bug -->
1.0
Convert JPEG file to PNG - # Description Script to convert JPEG file into PNG format. ## Type of issue - [x] Feature (New Script) ## Checklist: - [x] I have read the project guidelines. - [x] I have checked all the existing projects, before submitting a new project issue. - [x] I have checked previous issues to avoid duplicates. - [x] This issue will be meaningful for the project. <!-- Uncomment this in case you have a issue related to a bug in existing code.--> <!-- - [ ] I have added screenshots of the bug - [ ] I have added steps to reproduce the bug - [ ] I have proposed a possible solution for the bug -->
process
convert jpeg file to png description script to convert jpeg file into png format type of issue feature new script checklist i have read the project guidelines i have checked all the existing projects before submitting a new project issue i have checked previous issues to avoid duplicates this issue will be meaningful for the project i have added screenshots of the bug i have added steps to reproduce the bug i have proposed a possible solution for the bug
1
395,865
27,090,849,486
IssuesEvent
2023-02-14 20:55:58
larray-project/larray
https://api.github.com/repos/larray-project/larray
closed
update tutorial using get_example_filepath
bug documentation difficulty: low priority: BLOCKER
It now returns a Path object so this code is not valid anymore: ```python csv_dir = get_example_filepath('examples') population = read_csv(csv_dir + '/population.csv') ``` and should be changed to: ```python csv_dir = get_example_filepath('examples') population = read_csv(csv_dir / 'population.csv') ``` There might be some other usage outside of the tutorial.
1.0
update tutorial using get_example_filepath - It now returns a Path object so this code is not valid anymore: ```python csv_dir = get_example_filepath('examples') population = read_csv(csv_dir + '/population.csv') ``` and should be changed to: ```python csv_dir = get_example_filepath('examples') population = read_csv(csv_dir / 'population.csv') ``` There might be some other usage outside of the tutorial.
non_process
update tutorial using get example filepath it now returns a path object so this code is not valid anymore python csv dir get example filepath examples population read csv csv dir population csv and should be changed to python csv dir get example filepath examples population read csv csv dir population csv there might be some other usage outside of the tutorial
0
21,103
28,059,714,515
IssuesEvent
2023-03-29 11:50:22
NationalSecurityAgency/ghidra
https://api.github.com/repos/NationalSecurityAgency/ghidra
closed
MIPS64 ElfRelocationHandler issue
Feature: Loader/ELF Feature: Processor/MIPS Status: Internal
**Describe the bug** Attempting to analyze some MIPS64 object files with debug info, and I'm seeing several messages about failed relocations: `ERROR (ElfRelocationHandler) Failed to apply component-offset pointer at .debug_info::00000011` **To Reproduce** Steps to reproduce the behavior: 1. Grab an object file from any MIPS64 big endian toolchain (I used `/mips-img-elf/2017.10-07/mips-img-elf/lib/mips-r6-hard-newlib/lib64/libc.a/lib_a-fprintf.o` from [this one](http://codescape.mips.com/components/toolchain/2017.10-07/Codescape.GNU.Tools.Package.2017.10-07.for.MIPS.IMG.Bare.Metal.CentOS-5.x86_64.tar.gz)) a. This is a bit of a contrived example from a well-known library, I know, but I am not able to share the one I'm working with and this one seems to exhibit the same behavior 3. Load in Ghidra 4. Observe that DWARF information is not parsed into the Data Type Manager and the aforementioned error occurs **Expected behavior** DWARF info is parsed. **Environment (please complete the following information):** - OS: macOS 13.2.1 - Java Version: openjdk 17.0.6 2023-01-17 LTS - Ghidra Version: 10.2.3 - Ghidra Origin: Official GitHub distro **Additional context** Add any other context about the problem here.
1.0
MIPS64 ElfRelocationHandler issue - **Describe the bug** Attempting to analyze some MIPS64 object files with debug info, and I'm seeing several messages about failed relocations: `ERROR (ElfRelocationHandler) Failed to apply component-offset pointer at .debug_info::00000011` **To Reproduce** Steps to reproduce the behavior: 1. Grab an object file from any MIPS64 big endian toolchain (I used `/mips-img-elf/2017.10-07/mips-img-elf/lib/mips-r6-hard-newlib/lib64/libc.a/lib_a-fprintf.o` from [this one](http://codescape.mips.com/components/toolchain/2017.10-07/Codescape.GNU.Tools.Package.2017.10-07.for.MIPS.IMG.Bare.Metal.CentOS-5.x86_64.tar.gz)) a. This is a bit of a contrived example from a well-known library, I know, but I am not able to share the one I'm working with and this one seems to exhibit the same behavior 3. Load in Ghidra 4. Observe that DWARF information is not parsed into the Data Type Manager and the aforementioned error occurs **Expected behavior** DWARF info is parsed. **Environment (please complete the following information):** - OS: macOS 13.2.1 - Java Version: openjdk 17.0.6 2023-01-17 LTS - Ghidra Version: 10.2.3 - Ghidra Origin: Official GitHub distro **Additional context** Add any other context about the problem here.
process
elfrelocationhandler issue describe the bug attempting to analyze some object files with debug info and i m seeing several messages about failed relocations error elfrelocationhandler failed to apply component offset pointer at debug info to reproduce steps to reproduce the behavior grab an object file from any big endian toolchain i used mips img elf mips img elf lib mips hard newlib libc a lib a fprintf o from a this is a bit of a contrived example from a well known library i know but i am not able to share the one i m working with and this one seems to exhibit the same behavior load in ghidra observe that dwarf information is not parsed into the data type manager and the aforementioned error occurs expected behavior dwarf info is parsed environment please complete the following information os macos java version openjdk lts ghidra version ghidra origin official github distro additional context add any other context about the problem here
1
21,965
30,461,787,721
IssuesEvent
2023-07-17 07:31:35
zammad/zammad
https://api.github.com/repos/zammad/zammad
closed
Single select display value in trigger variable (#{ticket.abc.value}) is empty when "custom option sort" is used
bug trigger verified prioritised by payment object manager attribute variable processing
### Used Zammad Version 6.0. ### Environment - Installation method: any - Operating system: any - Database + version: [any - Elasticsearch version: any - Browser + version: MS Edge, v. 114.0.1823.67 (64-bit) ### Actual behaviour If a certain object is referenced via a value in a trigger, for example, and the referenced object has the checkbox for custom sorting marked, no value will be displayed in a ticket, neither as a note nor in an e-mail. ### Expected behaviour For example, if a particular object is referenced via a value in a trigger and the referenced object has the custom sort checkbox selected, the checkbox should not affect whether or not the trigger can provide a value, so the selected value should be displayed via the trigger by default (note, email.) ### Steps to reproduce the behaviour Create an object --> reference said object via the following variable #{ticket.xyz_one_two.value} --> create a trigger where the variable value is to display the value selected in a field. The trigger works as intended. However, when the user selects the custom sort option under the object, no value will be displayed. I am attaching a few screenshots: <img width="391" alt="Test1" src="https://github.com/zammad/zammad/assets/101350124/b892a008-ab39-432a-949c-4494bb02f954"> <img width="440" alt="Test2" src="https://github.com/zammad/zammad/assets/101350124/c6b09f60-6869-48ea-a7bc-013ad0fbe803"> ### Support Ticket Ticket#10132762 ### I'm sure this is a bug and no feature request or a general question. yes
1.0
Single select display value in trigger variable (#{ticket.abc.value}) is empty when "custom option sort" is used - ### Used Zammad Version 6.0. ### Environment - Installation method: any - Operating system: any - Database + version: [any - Elasticsearch version: any - Browser + version: MS Edge, v. 114.0.1823.67 (64-bit) ### Actual behaviour If a certain object is referenced via a value in a trigger, for example, and the referenced object has the checkbox for custom sorting marked, no value will be displayed in a ticket, neither as a note nor in an e-mail. ### Expected behaviour For example, if a particular object is referenced via a value in a trigger and the referenced object has the custom sort checkbox selected, the checkbox should not affect whether or not the trigger can provide a value, so the selected value should be displayed via the trigger by default (note, email.) ### Steps to reproduce the behaviour Create an object --> reference said object via the following variable #{ticket.xyz_one_two.value} --> create a trigger where the variable value is to display the value selected in a field. The trigger works as intended. However, when the user selects the custom sort option under the object, no value will be displayed. I am attaching a few screenshots: <img width="391" alt="Test1" src="https://github.com/zammad/zammad/assets/101350124/b892a008-ab39-432a-949c-4494bb02f954"> <img width="440" alt="Test2" src="https://github.com/zammad/zammad/assets/101350124/c6b09f60-6869-48ea-a7bc-013ad0fbe803"> ### Support Ticket Ticket#10132762 ### I'm sure this is a bug and no feature request or a general question. yes
process
single select display value in trigger variable ticket abc value is empty when custom option sort is used used zammad version environment installation method any operating system any database version any elasticsearch version any browser version ms edge v bit actual behaviour if a certain object is referenced via a value in a trigger for example and the referenced object has the checkbox for custom sorting marked no value will be displayed in a ticket neither as a note nor in an e mail expected behaviour for example if a particular object is referenced via a value in a trigger and the referenced object has the custom sort checkbox selected the checkbox should not affect whether or not the trigger can provide a value so the selected value should be displayed via the trigger by default note email steps to reproduce the behaviour create an object reference said object via the following variable ticket xyz one two value create a trigger where the variable value is to display the value selected in a field the trigger works as intended however when the user selects the custom sort option under the object no value will be displayed i am attaching a few screenshots img width alt src img width alt src support ticket ticket i m sure this is a bug and no feature request or a general question yes
1
2,539
5,300,272,015
IssuesEvent
2017-02-10 03:52:49
mitchellh/packer
https://api.github.com/repos/mitchellh/packer
closed
Packer hangs when compress post-processor runs out of disk space
bug post-processor/compress
Hey guys, perhaps an unusual case here - I ran out of disk on my macbook whilst compressing a vagrant box. The `packer` processes are left hung and I'm going to have to manually kill them. Pertinent packer output: ``` vmware-iso (vagrant): Compressing: Vagrantfile vmware-iso (vagrant): Compressing: disk-s001.vmdk vmware-iso (vagrant): Compressing: disk-s002.vmdk vmware-iso (vagrant): Compressing: disk-s003.vmdk vmware-iso (vagrant): Compressing: disk-s004.vmdk vmware-iso (vagrant): Compressing: disk-s005.vmdk ^C ^C ``` Some `ps`: ``` > ps -ef | grep packer 501 61889 14816 0 8:50am ttys001 0:04.88 packer build --only=vmware-iso -var iso_checksum=90b6d0431b13e880c904a16ac9fa42a1 -var iso_url=../../iso/OSX_InstallESD_10.10.5_14F27.dmg template.json 501 61890 61889 0 8:50am ttys001 0:01.83 /usr/local/bin/packer build --only=vmware-iso -var iso_checksum=90b6d0431b13e880c904a16ac9fa42a1 -var iso_url=../../iso/OSX_InstallESD_10.10.5_14F27.dmg template.json 501 61891 61890 0 8:50am ttys001 0:17.87 /usr/local/bin/packer-builder-vmware-iso 501 61892 61890 0 8:50am ttys001 0:00.09 /usr/local/bin/packer-provisioner-file 501 61893 61890 0 8:50am ttys001 0:00.76 /usr/local/bin/packer-provisioner-shell 501 61894 61890 0 8:50am ttys001 10:52.47 /usr/local/bin/packer-post-processor-vagrant 501 65007 54027 0 9:53am ttys003 0:00.00 grep --color=auto packer ```
1.0
Packer hangs when compress post-processor runs out of disk space - Hey guys, perhaps an unusual case here - I ran out of disk on my macbook whilst compressing a vagrant box. The `packer` processes are left hung and I'm going to have to manually kill them. Pertinent packer output: ``` vmware-iso (vagrant): Compressing: Vagrantfile vmware-iso (vagrant): Compressing: disk-s001.vmdk vmware-iso (vagrant): Compressing: disk-s002.vmdk vmware-iso (vagrant): Compressing: disk-s003.vmdk vmware-iso (vagrant): Compressing: disk-s004.vmdk vmware-iso (vagrant): Compressing: disk-s005.vmdk ^C ^C ``` Some `ps`: ``` > ps -ef | grep packer 501 61889 14816 0 8:50am ttys001 0:04.88 packer build --only=vmware-iso -var iso_checksum=90b6d0431b13e880c904a16ac9fa42a1 -var iso_url=../../iso/OSX_InstallESD_10.10.5_14F27.dmg template.json 501 61890 61889 0 8:50am ttys001 0:01.83 /usr/local/bin/packer build --only=vmware-iso -var iso_checksum=90b6d0431b13e880c904a16ac9fa42a1 -var iso_url=../../iso/OSX_InstallESD_10.10.5_14F27.dmg template.json 501 61891 61890 0 8:50am ttys001 0:17.87 /usr/local/bin/packer-builder-vmware-iso 501 61892 61890 0 8:50am ttys001 0:00.09 /usr/local/bin/packer-provisioner-file 501 61893 61890 0 8:50am ttys001 0:00.76 /usr/local/bin/packer-provisioner-shell 501 61894 61890 0 8:50am ttys001 10:52.47 /usr/local/bin/packer-post-processor-vagrant 501 65007 54027 0 9:53am ttys003 0:00.00 grep --color=auto packer ```
process
packer hangs when compress post processor runs out of disk space hey guys perhaps an unusual case here i ran out of disk on my macbook whilst compressing a vagrant box the packer processes are left hung and i m going to have to manually kill them pertinent packer output vmware iso vagrant compressing vagrantfile vmware iso vagrant compressing disk vmdk vmware iso vagrant compressing disk vmdk vmware iso vagrant compressing disk vmdk vmware iso vagrant compressing disk vmdk vmware iso vagrant compressing disk vmdk c c some ps ps ef grep packer packer build only vmware iso var iso checksum var iso url iso osx installesd dmg template json usr local bin packer build only vmware iso var iso checksum var iso url iso osx installesd dmg template json usr local bin packer builder vmware iso usr local bin packer provisioner file usr local bin packer provisioner shell usr local bin packer post processor vagrant grep color auto packer
1
21,816
30,316,622,237
IssuesEvent
2023-07-10 15:59:47
tdwg/dwc
https://api.github.com/repos/tdwg/dwc
closed
Change term - Add comment to namePublishedIn & namePublishedInID
Term - change Class - Taxon non-normative Process - complete
## Term change * Submitter: Markus Döring * Efficacy Justification (why is this change necessary?): The definition is potentially unclear for recombinations Current Term definition: https://dwc.tdwg.org/list/#dwc_namePublishedIn Current Term definition: https://dwc.tdwg.org/list/#dwc_namePublishedInID Proposed attributes of the new term version (Please put actual changes to be implemented in **bold** and ~strikethrough~): * Usage comments (recommendations regarding content, etc., not normative): **A citation of the first publication of the name in its given combination, not the basionym / original name. Recombinations are often not published in zoology, in which case namePublishedIn should be empty.** The same change is also needed for dwc:namePublishedInID
1.0
Change term - Add comment to namePublishedIn & namePublishedInID - ## Term change * Submitter: Markus Döring * Efficacy Justification (why is this change necessary?): The definition is potentially unclear for recombinations Current Term definition: https://dwc.tdwg.org/list/#dwc_namePublishedIn Current Term definition: https://dwc.tdwg.org/list/#dwc_namePublishedInID Proposed attributes of the new term version (Please put actual changes to be implemented in **bold** and ~strikethrough~): * Usage comments (recommendations regarding content, etc., not normative): **A citation of the first publication of the name in its given combination, not the basionym / original name. Recombinations are often not published in zoology, in which case namePublishedIn should be empty.** The same change is also needed for dwc:namePublishedInID
process
change term add comment to namepublishedin namepublishedinid term change submitter markus döring efficacy justification why is this change necessary the definition is potentially unclear for recombinations current term definition current term definition proposed attributes of the new term version please put actual changes to be implemented in bold and strikethrough usage comments recommendations regarding content etc not normative a citation of the first publication of the name in its given combination not the basionym original name recombinations are often not published in zoology in which case namepublishedin should be empty the same change is also needed for dwc namepublishedinid
1
2,578
5,341,874,555
IssuesEvent
2017-02-17 05:17:36
AllenFang/react-bootstrap-table
https://api.github.com/repos/AllenFang/react-bootstrap-table
reopened
Paging Dropdown CSS issue
enhancement inprocess
Hello, I have implemented nested grid functionality. Now, the paging dropdown of child grid is not opening correctly. Please check attached screenshot. Please do the needful ! Regards, Romil ![paging-dropdown](https://cloud.githubusercontent.com/assets/8926933/23011146/44273542-f445-11e6-9607-6ef2e5599c57.png)
1.0
Paging Dropdown CSS issue - Hello, I have implemented nested grid functionality. Now, the paging dropdown of child grid is not opening correctly. Please check attached screenshot. Please do the needful ! Regards, Romil ![paging-dropdown](https://cloud.githubusercontent.com/assets/8926933/23011146/44273542-f445-11e6-9607-6ef2e5599c57.png)
process
paging dropdown css issue hello i have implemented nested grid functionality now the paging dropdown of child grid is not opening correctly please check attached screenshot please do the needful regards romil
1
29,300
2,714,492,860
IssuesEvent
2015-04-10 04:26:44
AaronBuxbaum/TomatoTimer
https://api.github.com/repos/AaronBuxbaum/TomatoTimer
closed
Records/Analytics
enhancement high priority
The ability to save & analyze your data is pretty important to the Pomodoro Technique. This really has to be implemented For Records, I'm thinking during your break there should be a "did you finish the item? Yes/No"
1.0
Records/Analytics - The ability to save & analyze your data is pretty important to the Pomodoro Technique. This really has to be implemented For Records, I'm thinking during your break there should be a "did you finish the item? Yes/No"
non_process
records analytics the ability to save analyze your data is pretty important to the pomodoro technique this really has to be implemented for records i m thinking during your break there should be a did you finish the item yes no
0
18,787
24,691,391,319
IssuesEvent
2022-10-19 08:48:54
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[Mobile apps] Stats and Charts(Trends) functionality is not working
Bug P0 iOS Android Process: Fixed Process: Tested QA Process: Tested dev
AR: Stats and Charts(Trends) functionality is not working > Values are not getting displayed in the stats and charts when the participant provides the response for the activities ER: Stats and Chart functionality should work as expected
3.0
[Mobile apps] Stats and Charts(Trends) functionality is not working - AR: Stats and Charts(Trends) functionality is not working > Values are not getting displayed in the stats and charts when the participant provides the response for the activities ER: Stats and Chart functionality should work as expected
process
stats and charts trends functionality is not working ar stats and charts trends functionality is not working values are not getting displayed in the stats and charts when the participant provides the response for the activities er stats and chart functionality should work as expected
1
13,484
16,018,313,454
IssuesEvent
2021-04-20 18:57:42
googleapis/python-dialogflow
https://api.github.com/repos/googleapis/python-dialogflow
closed
How to select a region with the python client?
api: dialogflow type: process
Hi, guys, there is something wrong when I using the python client, because of dialogflow have different regions, every region has each agent, the client's default region is US, but when I want to select other region, I can't find any parameter and method helping me do it. Could you help me to solve it, thanks very much.
1.0
How to select a region with the python client? - Hi, guys, there is something wrong when I using the python client, because of dialogflow have different regions, every region has each agent, the client's default region is US, but when I want to select other region, I can't find any parameter and method helping me do it. Could you help me to solve it, thanks very much.
process
how to select a region with the python client hi guys there is something wrong when i using the python client because of dialogflow have different regions every region has each agent the client s default region is us but when i want to select other region i can t find any parameter and method helping me do it could you help me to solve it thanks very much
1
21,199
28,236,964,492
IssuesEvent
2023-04-06 02:00:09
lizhihao6/get-daily-arxiv-noti
https://api.github.com/repos/lizhihao6/get-daily-arxiv-noti
opened
New submissions for Thu, 6 Apr 23
event camera white balance isp compression image signal processing image signal process raw raw image events camera color contrast events AWB
## Keyword: events There is no result ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB ### SMPConv: Self-moving Point Representations for Continuous Convolution - **Authors:** Sanghyeon Kim, Eunbyung Park - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.02330 - **Pdf link:** https://arxiv.org/pdf/2304.02330 - **Abstract** Continuous convolution has recently gained prominence due to its ability to handle irregularly sampled data and model long-term dependency. Also, the promising experimental results of using large convolutional kernels have catalyzed the development of continuous convolution since they can construct large kernels very efficiently. Leveraging neural networks, more specifically multilayer perceptrons (MLPs), is by far the most prevalent approach to implementing continuous convolution. However, there are a few drawbacks, such as high computational costs, complex hyperparameter tuning, and limited descriptive power of filters. This paper suggests an alternative approach to building a continuous convolution without neural networks, resulting in more computationally efficient and improved performance. We present self-moving point representations where weight parameters freely move, and interpolation schemes are used to implement continuous functions. When applied to construct convolutional kernels, the experimental results have shown improved performance with drop-in replacement in the existing frameworks. Due to its lightweight structure, we are first to demonstrate the effectiveness of continuous convolution in a large-scale setting, e.g., ImageNet, presenting the improvements over the prior arts. Our code is available on https://github.com/sangnekim/SMPConv ## Keyword: ISP There is no result ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression ### HNeRV: A Hybrid Neural Representation for Videos - **Authors:** Hao Chen, Matt Gwilliam, Ser-Nam Lim, Abhinav Shrivastava - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.02633 - **Pdf link:** https://arxiv.org/pdf/2304.02633 - **Abstract** Implicit neural representations store videos as neural networks and have performed well for various vision tasks such as video compression and denoising. With frame index or positional index as input, implicit representations (NeRV, E-NeRV, \etc) reconstruct video from fixed and content-agnostic embeddings. Such embedding largely limits the regression capacity and internal generalization for video interpolation. In this paper, we propose a Hybrid Neural Representation for Videos (HNeRV), where a learnable encoder generates content-adaptive embeddings, which act as the decoder input. Besides the input embedding, we introduce HNeRV blocks, which ensure model parameters are evenly distributed across the entire network, such that higher layers (layers near the output) can have more capacity to store high-resolution content and video details. With content-adaptive embeddings and re-designed architecture, HNeRV outperforms implicit methods in video regression tasks for both reconstruction quality ($+4.7$ PSNR) and convergence speed ($16\times$ faster), and shows better internal generalization. As a simple and efficient video representation, HNeRV also shows decoding advantages for speed, flexibility, and deployment, compared to traditional codecs~(H.264, H.265) and learning-based compression methods. Finally, we explore the effectiveness of HNeRV on downstream tasks such as video compression and video inpainting. We provide project page at https://haochen-rye.github.io/HNeRV, and Code at https://github.com/haochen-rye/HNeRV ## Keyword: RAW ### A Diffusion-based Method for Multi-turn Compositional Image Generation - **Authors:** Chao Wang, Xiaoyu Yang, Jinmiao Huang, Kevin Ferreira - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2304.02192 - **Pdf link:** https://arxiv.org/pdf/2304.02192 - **Abstract** Multi-turn compositional image generation (M-CIG) is a challenging task that aims to iteratively manipulate a reference image given a modification text. While most of the existing methods for M-CIG are based on generative adversarial networks (GANs), recent advances in image generation have demonstrated the superiority of diffusion models over GANs. In this paper, we propose a diffusion-based method for M-CIG named conditional denoising diffusion with image compositional matching (CDD-ICM). We leverage CLIP as the backbone of image and text encoders, and incorporate a gated fusion mechanism, originally proposed for question answering, to compositionally fuse the reference image and the modification text at each turn of M-CIG. We introduce a conditioning scheme to generate the target image based on the fusion results. To prioritize the semantic quality of the generated target image, we learn an auxiliary image compositional match (ICM) objective, along with the conditional denoising diffusion (CDD) objective in a multi-task learning framework. Additionally, we also perform ICM guidance and classifier-free guidance to improve performance. Experimental results show that CDD-ICM achieves state-of-the-art results on two benchmark datasets for M-CIG, i.e., CoDraw and i-CLEVR. ### Multi-Domain Norm-referenced Encoding Enables Data Efficient Transfer Learning of Facial Expression Recognition - **Authors:** Michael Stettler, Alexander Lappe, Nick Taubert, Martin Giese - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.02309 - **Pdf link:** https://arxiv.org/pdf/2304.02309 - **Abstract** People can innately recognize human facial expressions in unnatural forms, such as when depicted on the unusual faces drawn in cartoons or when applied to an animal's features. However, current machine learning algorithms struggle with out-of-domain transfer in facial expression recognition (FER). We propose a biologically-inspired mechanism for such transfer learning, which is based on norm-referenced encoding, where patterns are encoded in terms of difference vectors relative to a domain-specific reference vector. By incorporating domain-specific reference frames, we demonstrate high data efficiency in transfer learning across multiple domains. Our proposed architecture provides an explanation for how the human brain might innately recognize facial expressions on varying head shapes (humans, monkeys, and cartoon avatars) without extensive training. Norm-referenced encoding also allows the intensity of the expression to be read out directly from neural unit activity, similar to face-selective neurons in the brain. Our model achieves a classification accuracy of 92.15\% on the FERG dataset with extreme data efficiency. We train our proposed mechanism with only 12 images, including a single image of each class (facial expression) and one image per domain (avatar). In comparison, the authors of the FERG dataset achieved a classification accuracy of 89.02\% with their FaceExpr model, which was trained on 43,000 images. ### SMPConv: Self-moving Point Representations for Continuous Convolution - **Authors:** Sanghyeon Kim, Eunbyung Park - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.02330 - **Pdf link:** https://arxiv.org/pdf/2304.02330 - **Abstract** Continuous convolution has recently gained prominence due to its ability to handle irregularly sampled data and model long-term dependency. Also, the promising experimental results of using large convolutional kernels have catalyzed the development of continuous convolution since they can construct large kernels very efficiently. Leveraging neural networks, more specifically multilayer perceptrons (MLPs), is by far the most prevalent approach to implementing continuous convolution. However, there are a few drawbacks, such as high computational costs, complex hyperparameter tuning, and limited descriptive power of filters. This paper suggests an alternative approach to building a continuous convolution without neural networks, resulting in more computationally efficient and improved performance. We present self-moving point representations where weight parameters freely move, and interpolation schemes are used to implement continuous functions. When applied to construct convolutional kernels, the experimental results have shown improved performance with drop-in replacement in the existing frameworks. Due to its lightweight structure, we are first to demonstrate the effectiveness of continuous convolution in a large-scale setting, e.g., ImageNet, presenting the improvements over the prior arts. Our code is available on https://github.com/sangnekim/SMPConv ## Keyword: raw image There is no result
2.0
New submissions for Thu, 6 Apr 23 - ## Keyword: events There is no result ## Keyword: event camera There is no result ## Keyword: events camera There is no result ## Keyword: white balance There is no result ## Keyword: color contrast There is no result ## Keyword: AWB ### SMPConv: Self-moving Point Representations for Continuous Convolution - **Authors:** Sanghyeon Kim, Eunbyung Park - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.02330 - **Pdf link:** https://arxiv.org/pdf/2304.02330 - **Abstract** Continuous convolution has recently gained prominence due to its ability to handle irregularly sampled data and model long-term dependency. Also, the promising experimental results of using large convolutional kernels have catalyzed the development of continuous convolution since they can construct large kernels very efficiently. Leveraging neural networks, more specifically multilayer perceptrons (MLPs), is by far the most prevalent approach to implementing continuous convolution. However, there are a few drawbacks, such as high computational costs, complex hyperparameter tuning, and limited descriptive power of filters. This paper suggests an alternative approach to building a continuous convolution without neural networks, resulting in more computationally efficient and improved performance. We present self-moving point representations where weight parameters freely move, and interpolation schemes are used to implement continuous functions. When applied to construct convolutional kernels, the experimental results have shown improved performance with drop-in replacement in the existing frameworks. Due to its lightweight structure, we are first to demonstrate the effectiveness of continuous convolution in a large-scale setting, e.g., ImageNet, presenting the improvements over the prior arts. Our code is available on https://github.com/sangnekim/SMPConv ## Keyword: ISP There is no result ## Keyword: image signal processing There is no result ## Keyword: image signal process There is no result ## Keyword: compression ### HNeRV: A Hybrid Neural Representation for Videos - **Authors:** Hao Chen, Matt Gwilliam, Ser-Nam Lim, Abhinav Shrivastava - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.02633 - **Pdf link:** https://arxiv.org/pdf/2304.02633 - **Abstract** Implicit neural representations store videos as neural networks and have performed well for various vision tasks such as video compression and denoising. With frame index or positional index as input, implicit representations (NeRV, E-NeRV, \etc) reconstruct video from fixed and content-agnostic embeddings. Such embedding largely limits the regression capacity and internal generalization for video interpolation. In this paper, we propose a Hybrid Neural Representation for Videos (HNeRV), where a learnable encoder generates content-adaptive embeddings, which act as the decoder input. Besides the input embedding, we introduce HNeRV blocks, which ensure model parameters are evenly distributed across the entire network, such that higher layers (layers near the output) can have more capacity to store high-resolution content and video details. With content-adaptive embeddings and re-designed architecture, HNeRV outperforms implicit methods in video regression tasks for both reconstruction quality ($+4.7$ PSNR) and convergence speed ($16\times$ faster), and shows better internal generalization. As a simple and efficient video representation, HNeRV also shows decoding advantages for speed, flexibility, and deployment, compared to traditional codecs~(H.264, H.265) and learning-based compression methods. Finally, we explore the effectiveness of HNeRV on downstream tasks such as video compression and video inpainting. We provide project page at https://haochen-rye.github.io/HNeRV, and Code at https://github.com/haochen-rye/HNeRV ## Keyword: RAW ### A Diffusion-based Method for Multi-turn Compositional Image Generation - **Authors:** Chao Wang, Xiaoyu Yang, Jinmiao Huang, Kevin Ferreira - **Subjects:** Computer Vision and Pattern Recognition (cs.CV); Artificial Intelligence (cs.AI); Machine Learning (cs.LG) - **Arxiv link:** https://arxiv.org/abs/2304.02192 - **Pdf link:** https://arxiv.org/pdf/2304.02192 - **Abstract** Multi-turn compositional image generation (M-CIG) is a challenging task that aims to iteratively manipulate a reference image given a modification text. While most of the existing methods for M-CIG are based on generative adversarial networks (GANs), recent advances in image generation have demonstrated the superiority of diffusion models over GANs. In this paper, we propose a diffusion-based method for M-CIG named conditional denoising diffusion with image compositional matching (CDD-ICM). We leverage CLIP as the backbone of image and text encoders, and incorporate a gated fusion mechanism, originally proposed for question answering, to compositionally fuse the reference image and the modification text at each turn of M-CIG. We introduce a conditioning scheme to generate the target image based on the fusion results. To prioritize the semantic quality of the generated target image, we learn an auxiliary image compositional match (ICM) objective, along with the conditional denoising diffusion (CDD) objective in a multi-task learning framework. Additionally, we also perform ICM guidance and classifier-free guidance to improve performance. Experimental results show that CDD-ICM achieves state-of-the-art results on two benchmark datasets for M-CIG, i.e., CoDraw and i-CLEVR. ### Multi-Domain Norm-referenced Encoding Enables Data Efficient Transfer Learning of Facial Expression Recognition - **Authors:** Michael Stettler, Alexander Lappe, Nick Taubert, Martin Giese - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.02309 - **Pdf link:** https://arxiv.org/pdf/2304.02309 - **Abstract** People can innately recognize human facial expressions in unnatural forms, such as when depicted on the unusual faces drawn in cartoons or when applied to an animal's features. However, current machine learning algorithms struggle with out-of-domain transfer in facial expression recognition (FER). We propose a biologically-inspired mechanism for such transfer learning, which is based on norm-referenced encoding, where patterns are encoded in terms of difference vectors relative to a domain-specific reference vector. By incorporating domain-specific reference frames, we demonstrate high data efficiency in transfer learning across multiple domains. Our proposed architecture provides an explanation for how the human brain might innately recognize facial expressions on varying head shapes (humans, monkeys, and cartoon avatars) without extensive training. Norm-referenced encoding also allows the intensity of the expression to be read out directly from neural unit activity, similar to face-selective neurons in the brain. Our model achieves a classification accuracy of 92.15\% on the FERG dataset with extreme data efficiency. We train our proposed mechanism with only 12 images, including a single image of each class (facial expression) and one image per domain (avatar). In comparison, the authors of the FERG dataset achieved a classification accuracy of 89.02\% with their FaceExpr model, which was trained on 43,000 images. ### SMPConv: Self-moving Point Representations for Continuous Convolution - **Authors:** Sanghyeon Kim, Eunbyung Park - **Subjects:** Computer Vision and Pattern Recognition (cs.CV) - **Arxiv link:** https://arxiv.org/abs/2304.02330 - **Pdf link:** https://arxiv.org/pdf/2304.02330 - **Abstract** Continuous convolution has recently gained prominence due to its ability to handle irregularly sampled data and model long-term dependency. Also, the promising experimental results of using large convolutional kernels have catalyzed the development of continuous convolution since they can construct large kernels very efficiently. Leveraging neural networks, more specifically multilayer perceptrons (MLPs), is by far the most prevalent approach to implementing continuous convolution. However, there are a few drawbacks, such as high computational costs, complex hyperparameter tuning, and limited descriptive power of filters. This paper suggests an alternative approach to building a continuous convolution without neural networks, resulting in more computationally efficient and improved performance. We present self-moving point representations where weight parameters freely move, and interpolation schemes are used to implement continuous functions. When applied to construct convolutional kernels, the experimental results have shown improved performance with drop-in replacement in the existing frameworks. Due to its lightweight structure, we are first to demonstrate the effectiveness of continuous convolution in a large-scale setting, e.g., ImageNet, presenting the improvements over the prior arts. Our code is available on https://github.com/sangnekim/SMPConv ## Keyword: raw image There is no result
process
new submissions for thu apr keyword events there is no result keyword event camera there is no result keyword events camera there is no result keyword white balance there is no result keyword color contrast there is no result keyword awb smpconv self moving point representations for continuous convolution authors sanghyeon kim eunbyung park subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract continuous convolution has recently gained prominence due to its ability to handle irregularly sampled data and model long term dependency also the promising experimental results of using large convolutional kernels have catalyzed the development of continuous convolution since they can construct large kernels very efficiently leveraging neural networks more specifically multilayer perceptrons mlps is by far the most prevalent approach to implementing continuous convolution however there are a few drawbacks such as high computational costs complex hyperparameter tuning and limited descriptive power of filters this paper suggests an alternative approach to building a continuous convolution without neural networks resulting in more computationally efficient and improved performance we present self moving point representations where weight parameters freely move and interpolation schemes are used to implement continuous functions when applied to construct convolutional kernels the experimental results have shown improved performance with drop in replacement in the existing frameworks due to its lightweight structure we are first to demonstrate the effectiveness of continuous convolution in a large scale setting e g imagenet presenting the improvements over the prior arts our code is available on keyword isp there is no result keyword image signal processing there is no result keyword image signal process there is no result keyword compression hnerv a hybrid neural representation for videos authors hao chen matt gwilliam ser nam lim abhinav shrivastava subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract implicit neural representations store videos as neural networks and have performed well for various vision tasks such as video compression and denoising with frame index or positional index as input implicit representations nerv e nerv etc reconstruct video from fixed and content agnostic embeddings such embedding largely limits the regression capacity and internal generalization for video interpolation in this paper we propose a hybrid neural representation for videos hnerv where a learnable encoder generates content adaptive embeddings which act as the decoder input besides the input embedding we introduce hnerv blocks which ensure model parameters are evenly distributed across the entire network such that higher layers layers near the output can have more capacity to store high resolution content and video details with content adaptive embeddings and re designed architecture hnerv outperforms implicit methods in video regression tasks for both reconstruction quality psnr and convergence speed times faster and shows better internal generalization as a simple and efficient video representation hnerv also shows decoding advantages for speed flexibility and deployment compared to traditional codecs h h and learning based compression methods finally we explore the effectiveness of hnerv on downstream tasks such as video compression and video inpainting we provide project page at and code at keyword raw a diffusion based method for multi turn compositional image generation authors chao wang xiaoyu yang jinmiao huang kevin ferreira subjects computer vision and pattern recognition cs cv artificial intelligence cs ai machine learning cs lg arxiv link pdf link abstract multi turn compositional image generation m cig is a challenging task that aims to iteratively manipulate a reference image given a modification text while most of the existing methods for m cig are based on generative adversarial networks gans recent advances in image generation have demonstrated the superiority of diffusion models over gans in this paper we propose a diffusion based method for m cig named conditional denoising diffusion with image compositional matching cdd icm we leverage clip as the backbone of image and text encoders and incorporate a gated fusion mechanism originally proposed for question answering to compositionally fuse the reference image and the modification text at each turn of m cig we introduce a conditioning scheme to generate the target image based on the fusion results to prioritize the semantic quality of the generated target image we learn an auxiliary image compositional match icm objective along with the conditional denoising diffusion cdd objective in a multi task learning framework additionally we also perform icm guidance and classifier free guidance to improve performance experimental results show that cdd icm achieves state of the art results on two benchmark datasets for m cig i e codraw and i clevr multi domain norm referenced encoding enables data efficient transfer learning of facial expression recognition authors michael stettler alexander lappe nick taubert martin giese subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract people can innately recognize human facial expressions in unnatural forms such as when depicted on the unusual faces drawn in cartoons or when applied to an animal s features however current machine learning algorithms struggle with out of domain transfer in facial expression recognition fer we propose a biologically inspired mechanism for such transfer learning which is based on norm referenced encoding where patterns are encoded in terms of difference vectors relative to a domain specific reference vector by incorporating domain specific reference frames we demonstrate high data efficiency in transfer learning across multiple domains our proposed architecture provides an explanation for how the human brain might innately recognize facial expressions on varying head shapes humans monkeys and cartoon avatars without extensive training norm referenced encoding also allows the intensity of the expression to be read out directly from neural unit activity similar to face selective neurons in the brain our model achieves a classification accuracy of on the ferg dataset with extreme data efficiency we train our proposed mechanism with only images including a single image of each class facial expression and one image per domain avatar in comparison the authors of the ferg dataset achieved a classification accuracy of with their faceexpr model which was trained on images smpconv self moving point representations for continuous convolution authors sanghyeon kim eunbyung park subjects computer vision and pattern recognition cs cv arxiv link pdf link abstract continuous convolution has recently gained prominence due to its ability to handle irregularly sampled data and model long term dependency also the promising experimental results of using large convolutional kernels have catalyzed the development of continuous convolution since they can construct large kernels very efficiently leveraging neural networks more specifically multilayer perceptrons mlps is by far the most prevalent approach to implementing continuous convolution however there are a few drawbacks such as high computational costs complex hyperparameter tuning and limited descriptive power of filters this paper suggests an alternative approach to building a continuous convolution without neural networks resulting in more computationally efficient and improved performance we present self moving point representations where weight parameters freely move and interpolation schemes are used to implement continuous functions when applied to construct convolutional kernels the experimental results have shown improved performance with drop in replacement in the existing frameworks due to its lightweight structure we are first to demonstrate the effectiveness of continuous convolution in a large scale setting e g imagenet presenting the improvements over the prior arts our code is available on keyword raw image there is no result
1
327,474
9,976,159,600
IssuesEvent
2019-07-09 14:31:45
OpenNebula/one
https://api.github.com/repos/OpenNebula/one
opened
XSD file for zone should match with the output of one.zone.info
Category: API Priority: Normal Status: Accepted
**Description** XSD file for zone should match with the output of `one.zone.info` instead of `onezone show <zone_id> -x`. Also it is necessary to create a raftstatus.xsd file for defining the strucutre of the output of `one.zone.raftstatus`. <!--////////////////////////////////////////////--> <!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM --> <!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS --> <!-- PROGRESS WILL BE REFLECTED HERE --> <!--////////////////////////////////////////////--> ## Progress Status - [ ] Branch created - [ ] Code committed to development branch - [ ] Testing - QA - [ ] Documentation - [ ] Release notes - resolved issues, compatibility, known issues - [ ] Code committed to upstream release/hotfix branches - [ ] Documentation committed to upstream release/hotfix branches
1.0
XSD file for zone should match with the output of one.zone.info - **Description** XSD file for zone should match with the output of `one.zone.info` instead of `onezone show <zone_id> -x`. Also it is necessary to create a raftstatus.xsd file for defining the strucutre of the output of `one.zone.raftstatus`. <!--////////////////////////////////////////////--> <!-- THIS SECTION IS FOR THE DEVELOPMENT TEAM --> <!-- BOTH FOR BUGS AND ENHANCEMENT REQUESTS --> <!-- PROGRESS WILL BE REFLECTED HERE --> <!--////////////////////////////////////////////--> ## Progress Status - [ ] Branch created - [ ] Code committed to development branch - [ ] Testing - QA - [ ] Documentation - [ ] Release notes - resolved issues, compatibility, known issues - [ ] Code committed to upstream release/hotfix branches - [ ] Documentation committed to upstream release/hotfix branches
non_process
xsd file for zone should match with the output of one zone info description xsd file for zone should match with the output of one zone info instead of onezone show x also it is necessary to create a raftstatus xsd file for defining the strucutre of the output of one zone raftstatus progress status branch created code committed to development branch testing qa documentation release notes resolved issues compatibility known issues code committed to upstream release hotfix branches documentation committed to upstream release hotfix branches
0
311,265
23,378,560,019
IssuesEvent
2022-08-11 07:11:23
singularity-data/risingwave-docs
https://api.github.com/repos/singularity-data/risingwave-docs
opened
Document table functions: unnest(), generate_series
documentation
### Related code PR https://github.com/singularity-data/risingwave/pull/4030 ### Which part(s) of the docs might be affected or should be updated? And how? Support table functions in SELECT. Add a new section, "Table functions" to <img width="276" alt="image" src="https://user-images.githubusercontent.com/100549427/184081146-038cf7ed-b721-4f54-8bc4-9032efeb6e74.png"> Add unnest() and generate_series to "Table functions". Maybe add relevant examples to [SELECT](https://www.risingwave.dev/docs/latest/sql-select/). ### Reference https://www.postgresql.org/docs/14/queries-table-expressions.html https://www.postgresql.org/docs/14/functions-array.html https://count.co/sql-resources/bigquery-standard-sql/unnest
1.0
Document table functions: unnest(), generate_series - ### Related code PR https://github.com/singularity-data/risingwave/pull/4030 ### Which part(s) of the docs might be affected or should be updated? And how? Support table functions in SELECT. Add a new section, "Table functions" to <img width="276" alt="image" src="https://user-images.githubusercontent.com/100549427/184081146-038cf7ed-b721-4f54-8bc4-9032efeb6e74.png"> Add unnest() and generate_series to "Table functions". Maybe add relevant examples to [SELECT](https://www.risingwave.dev/docs/latest/sql-select/). ### Reference https://www.postgresql.org/docs/14/queries-table-expressions.html https://www.postgresql.org/docs/14/functions-array.html https://count.co/sql-resources/bigquery-standard-sql/unnest
non_process
document table functions unnest generate series related code pr which part s of the docs might be affected or should be updated and how support table functions in select add a new section table functions to img width alt image src add unnest and generate series to table functions maybe add relevant examples to reference
0
2,331
5,142,627,370
IssuesEvent
2017-01-12 13:53:54
jimbrown75/Permit-Vision-Enhancements
https://api.github.com/repos/jimbrown75/Permit-Vision-Enhancements
opened
2D Barcodes
bug High Priority Must Fix Verified by PTW Process Lead
Defect - The existing isolation tag barcodes do not scan from a mobile device. eVision advise that 2D barcodes are required. Please include in V8 MD delivery. 2D barcodes must be available for permits, ICC's and ICC tags.
1.0
2D Barcodes - Defect - The existing isolation tag barcodes do not scan from a mobile device. eVision advise that 2D barcodes are required. Please include in V8 MD delivery. 2D barcodes must be available for permits, ICC's and ICC tags.
process
barcodes defect the existing isolation tag barcodes do not scan from a mobile device evision advise that barcodes are required please include in md delivery barcodes must be available for permits icc s and icc tags
1
76
2,528,310,320
IssuesEvent
2015-01-22 01:36:01
tinkerpop/tinkerpop3
https://api.github.com/repos/tinkerpop/tinkerpop3
closed
We may be able to get rid of Step.setNextStep() and setPreviousStep()....
enhancement process
...by means of being smart about step indexing. getNextStep() and getPreviousStep() will still exist.
1.0
We may be able to get rid of Step.setNextStep() and setPreviousStep().... - ...by means of being smart about step indexing. getNextStep() and getPreviousStep() will still exist.
process
we may be able to get rid of step setnextstep and setpreviousstep by means of being smart about step indexing getnextstep and getpreviousstep will still exist
1
20,041
26,525,796,214
IssuesEvent
2023-01-19 08:37:17
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
regulation of endosome size (GO:0051036)
obsoletion waiting for feedback cellular processes
Please provide as much information as you can: * **GO term ID and Label** regulation of endosome size (GO:0051036) * **Reason for deprecation** This seems like a phenotype, not a process. Endosome size changes in different mutants as steps in endosomal trafficking are blocked. Size of endosomes is not regulated per. se. most likely just "endocytosis" or "regulation of endocytosis" * **Are there annotations to this term?** - How many EXP: 31 + | − | (23) | UniProt + | − | (5) | FlyBase + | − | (2) | RGD + | − | (1) | SGD * **Are there mappings and cross references to this term? (InterPro, Keywords; check QuickGO cross-references section)** * **Is this term in a subset? (check the AmiGO page for that term)** * **Any other information**
1.0
regulation of endosome size (GO:0051036) - Please provide as much information as you can: * **GO term ID and Label** regulation of endosome size (GO:0051036) * **Reason for deprecation** This seems like a phenotype, not a process. Endosome size changes in different mutants as steps in endosomal trafficking are blocked. Size of endosomes is not regulated per. se. most likely just "endocytosis" or "regulation of endocytosis" * **Are there annotations to this term?** - How many EXP: 31 + | − | (23) | UniProt + | − | (5) | FlyBase + | − | (2) | RGD + | − | (1) | SGD * **Are there mappings and cross references to this term? (InterPro, Keywords; check QuickGO cross-references section)** * **Is this term in a subset? (check the AmiGO page for that term)** * **Any other information**
process
regulation of endosome size go please provide as much information as you can go term id and label regulation of endosome size go reason for deprecation this seems like a phenotype not a process endosome size changes in different mutants as steps in endosomal trafficking are blocked size of endosomes is not regulated per se most likely just endocytosis or regulation of endocytosis are there annotations to this term how many exp − uniprot − flybase − rgd − sgd are there mappings and cross references to this term interpro keywords check quickgo cross references section is this term in a subset check the amigo page for that term any other information
1
216,812
16,819,953,424
IssuesEvent
2021-06-17 11:58:44
eclipse-openj9/openj9
https://api.github.com/repos/eclipse-openj9/openj9
opened
JDK8 MacOS MauveMultiThrdLoad_special_5m_22_FAILED - Segmentation error vmState=0x00000000 Module=default/libj9jit29.dylib
test failure
Failure link ------------ From an internal build `Test_openjdk8_j9_special.system_x86-64_mac_testList_0/44/` (`osxrt11`): ``` openjdk version "1.8.0_302-beta" OpenJDK Runtime Environment (build 1.8.0_302-beta-202106170233-b05) Eclipse OpenJ9 VM (build master-0385f0737, JRE 1.8.0 Mac OS X amd64-64-Bit Compressed References 20210616_155 (JIT enabled, AOT enabled) OpenJ9 - 0385f0737 OMR - 07e7658a9 JCL - be819230 based on jdk8u302-b05) ``` [Rerun in Grinder](https://hyc-runtimes-jenkins.swg-devops.com/job/Grinder/parambuild/?SDK_RESOURCE=upstream&amp;TARGET=MauveMultiThrdLoad_special_5m_22&amp;TEST_FLAG=&amp;UPSTREAM_TEST_JOB_NAME=Test_openjdk8_j9_special.system_x86-64_mac&amp;DOCKER_REQUIRED=false&amp;ACTIVE_NODE_TIMEOUT=0&amp;VENDOR_TEST_DIRS=&amp;EXTRA_DOCKER_ARGS=&amp;TKG_OWNER_BRANCH=AdoptOpenJDK%3Amaster&amp;TEST_PARALLELIZATION_PARAMS=&amp;OPENJ9_SYSTEMTEST_OWNER_BRANCH=eclipse%3Amaster&amp;PLATFORM=x86-64_mac&amp;GENERATE_JOBS=true&amp;KEEP_REPORTDIR=false&amp;PERSONAL_BUILD=false&amp;ADOPTOPENJDK_REPO=https%3A%2F%2Fgithub.com%2Fadoptium%2Faqa-tests.git&amp;LABEL=&amp;TEST_OPTIONS_PARAMS=&amp;EXTRA_OPTIONS=&amp;CUSTOMIZED_SDK_URL=+https%3A%2F%2Fna.artifactory.swg-devops.com%2Fartifactory%2Fsys-rt-generic-local%2Fhyc-runtimes-jenkins.swg-devops.com%2Fbuild-scripts%2Fjobs%2Fjdk8u%2Fjdk8u-mac-x64-openj9%2F155%2FOpenJDK8U-jdk_x64_mac_openj9_2021-06-17-02-30.tar.gz+https%3A%2F%2Fna.artifactory.swg-devops.com%2Fartifactory%2Fsys-rt-generic-local%2Fhyc-runtimes-jenkins.swg-devops.com%2Fbuild-scripts%2Fjobs%2Fjdk8u%2Fjdk8u-mac-x64-openj9%2F155%2FOpenJDK8U-jre_x64_mac_openj9_2021-06-17-02-30.tar.gz+https%3A%2F%2Fna.artifactory.swg-devops.com%2Fartifactory%2Fsys-rt-generic-local%2Fhyc-runtimes-jenkins.swg-devops.com%2Fbuild-scripts%2Fjobs%2Fjdk8u%2Fjdk8u-mac-x64-openj9%2F155%2FOpenJDK8U-testimage_x64_mac_openj9_2021-06-17-02-30.tar.gz+https%3A%2F%2Fna.artifactory.swg-devops.com%2Fartifactory%2Fsys-rt-generic-local%2Fhyc-runtimes-jenkins.swg-devops.com%2Fbuild-scripts%2Fjobs%2Fjdk8u%2Fjdk8u-mac-x64-openj9%2F155%2FOpenJDK8U-debugimage_x64_mac_openj9_2021-06-17-02-30.tar.gz&amp;BUILD_IDENTIFIER=&amp;NON_AQA_TEST_REPOS_HELP_TEXT=&amp;ADOPTOPENJDK_BRANCH=master&amp;LIGHT_WEIGHT_CHECKOUT=true&amp;NON_AQA_TEST_REPOS=&amp;ARTIFACTORY_SERVER=na.artifactory.swg-devops&amp;TEST_REPO_PARAMS=&amp;TEST_SELECTION_PARAMS=&amp;TEST_PARALLELIZATION_PARAMS_HELP_TEXT=&amp;JDK_SELECTION_PARAMS=&amp;KEEP_WORKSPACE=false&amp;USER_CREDENTIALS_ID=&amp;JDK_VERSION=8&amp;ITERATIONS=1&amp;VENDOR_TEST_REPOS=&amp;JDK_REPO=https%3A%2F%2Fgithub.com%2Fibmruntimes%2Fopenj9-openjdk-jdk8&amp;PLATFORM_AND_MACHINE_HELP_TEXT=&amp;RELEASE_TAG=&amp;OPENJ9_BRANCH=master&amp;OPENJ9_SHA=&amp;JCK_GIT_REPO=&amp;VENDOR_TEST_BRANCHES=&amp;OPENJ9_REPO=https%3A%2F%2Fgithub.com%2Feclipse-openj9%2Fopenj9.git&amp;UPSTREAM_JOB_NAME=&amp;CLOUD_PROVIDER=&amp;PLATFORM_AND_MACHINE=&amp;CUSTOM_TARGET=&amp;VENDOR_TEST_SHAS=&amp;JDK_BRANCH=openj9&amp;LABEL_ADDITION=&amp;ARTIFACTORY_REPO=&amp;ARTIFACTORY_ROOT_DIR=&amp;POST_RUN_PARAMS_HELP_TEXT=&amp;UPSTREAM_TEST_JOB_NUMBER=45&amp;DOCKERIMAGE_TAG=&amp;JDK_SELECTION_PARAMS_HELP_TEXT=&amp;JDK_IMPL=openj9&amp;TEST_TIME=&amp;SSH_AGENT_CREDENTIAL=83181e25-eea4-4f55-8b3e-e79615733226&amp;AUTO_DETECT=true&amp;TKG_SHA=&amp;TEST_SELECTION_PARAMS_HELP_TEXT=&amp;ADOPTOPENJDK_SYSTEMTEST_OWNER_BRANCH=adoptium%3Amaster&amp;CUSTOMIZED_SDK_URL_CREDENTIAL_ID=4e18ffe7-b1b1-4272-9979-99769b68bcc2&amp;OPENJDK_SHA=&amp;NUM_MACHINES=&amp;TRSS_URL=http%3A%2F%2Ftrss1.fyre.ibm.com&amp;BUILD_LIST=system&amp;UPSTREAM_JOB_NUMBER=&amp;STF_OWNER_BRANCH=adoptium%3Amaster&amp;TEST_REPO_PARAMS_HELP_TEXT=&amp;POST_RUN_PARAMS=&amp;TIME_LIMIT=20&amp;TEST_OPTIONS_PARAMS_HELP_TEXT=&amp;JVM_OPTIONS=&amp;PARALLEL=None) Optional info ------------- Failure output (captured from console output) --------------------------------------------- ``` [2021-06-17T08:37:05.525Z] Running test MauveMultiThrdLoad_special_5m_22 ... [2021-06-17T08:37:05.525Z] =============================================== [2021-06-17T08:37:05.525Z] MauveMultiThrdLoad_special_5m_22 Start Time: Thu Jun 17 04:37:04 2021 Epoch Time (ms): 1623919024447 [2021-06-17T08:37:05.525Z] "/Users/jenkins/workspace/Test_openjdk8_j9_special.system_x86-64_mac_testList_0/openjdkbinary/j2sdk-image/Contents/Home/bin/..//bin/java" -Xshareclasses:destroyAll; "/Users/jenkins/workspace/Test_openjdk8_j9_special.system_x86-64_mac_testList_0/openjdkbinary/j2sdk-image/Contents/Home/bin/..//bin/java" -Xshareclasses:groupAccess,destroyAll; echo "cache cleanup done"; [2021-06-17T08:37:05.525Z] JVMSHRC005I No shared class caches available [2021-06-17T08:37:05.525Z] JVMSHRC005I No shared class caches available [2021-06-17T08:37:05.525Z] cache cleanup done [2021-06-17T08:37:05.525Z] variation: Mode688 [2021-06-17T08:37:05.525Z] JVM_OPTIONS: -Xcompressedrefs -Xjit:count=0 -Xgcpolicy:gencon -Xaggressive -Xconcurrentlevel0 [2021-06-17T08:37:19.574Z] LT stderr Unhandled exception [2021-06-17T08:37:19.574Z] LT stderr Type=Segmentation error vmState=0x00000000 [2021-06-17T08:37:19.574Z] LT stderr J9Generic_Signal_Number=00000018 Signal_Number=0000000b Error_Value=00000000 Signal_Code=00000001 [2021-06-17T08:37:19.574Z] LT stderr Handler1=000000000D236CE0 Handler2=000000000D510C10 InaccessibleAddress=00000000FFE398B8 [2021-06-17T08:37:19.574Z] LT stderr RDI=000000002351A2DB RSI=000000002351A2E0 RAX=0000000000000048 RBX=0000000016883300 [2021-06-17T08:37:19.574Z] LT stderr RCX=00000000FFE39700 RDX=000000002351A2E0 R8=0000000016883638 R9=00000007FFE39410 [2021-06-17T08:37:19.574Z] LT stderr R10=0000000700099F80 R11=000000000B9EE050 R12=00000007FFCB1760 R13=00000007FFCB2D88 [2021-06-17T08:37:19.574Z] LT stderr R14=000000000000000A R15=00000007FFCB2D30 [2021-06-17T08:37:19.574Z] LT stderr RIP=000000000D9BB4D0 GS=0000 FS=0000 RSP=00000000168592D0 [2021-06-17T08:37:19.574Z] LT stderr RFlags=0000000000010217 CS=002B RBP=0000000016823300 ERR=FFE398B800000004 [2021-06-17T08:37:19.574Z] LT stderr TRAPNO=000000040000000E CPU=98B8000000040000 FAULTVADDR=00000000FFE398B8 [2021-06-17T08:37:19.574Z] LT stderr XMM0 0000000000000002 (f: 2.000000, d: 9.881313e-324) [2021-06-17T08:37:19.574Z] LT stderr XMM1 0000000016859318 (f: 377852704.000000, d: 1.866840e-315) [2021-06-17T08:37:19.574Z] LT stderr XMM2 0000000000000000 (f: 0.000000, d: 0.000000e+00) [2021-06-17T08:37:19.574Z] LT stderr XMM3 0000000000000000 (f: 0.000000, d: 0.000000e+00) [2021-06-17T08:37:19.574Z] LT stderr XMM4 0000000000000000 (f: 0.000000, d: 0.000000e+00) [2021-06-17T08:37:19.574Z] LT stderr XMM5 00000000001c0000 (f: 1835008.000000, d: 9.066144e-318) [2021-06-17T08:37:19.574Z] LT stderr XMM6 3fd89a3386c1425b (f: 2260812288.000000, d: 3.844117e-01) [2021-06-17T08:37:19.574Z] LT stderr XMM7 401e7f9c1e980fa9 (f: 513281952.000000, d: 7.624619e+00) [2021-06-17T08:37:19.574Z] LT stderr XMM8 ae5ff75e430f4000 (f: 1125072896.000000, d: -2.571067e-85) [2021-06-17T08:37:19.574Z] LT stderr XMM9 0000000000000000 (f: 0.000000, d: 0.000000e+00) [2021-06-17T08:37:19.574Z] LT stderr XMM10 0000000000000006 (f: 6.000000, d: 2.964394e-323) [2021-06-17T08:37:19.574Z] LT stderr XMM11 0000000000000008 (f: 8.000000, d: 3.952525e-323) [2021-06-17T08:37:19.574Z] LT stderr XMM12 0000000000000000 (f: 0.000000, d: 0.000000e+00) [2021-06-17T08:37:19.574Z] LT stderr XMM13 0000000000000000 (f: 0.000000, d: 0.000000e+00) [2021-06-17T08:37:19.574Z] LT stderr XMM14 0000000000000000 (f: 0.000000, d: 0.000000e+00) [2021-06-17T08:37:19.574Z] LT stderr XMM15 0000000000000000 (f: 0.000000, d: 0.000000e+00) [2021-06-17T08:37:19.574Z] LT stderr Module=/Users/jenkins/workspace/Test_openjdk8_j9_special.system_x86-64_mac_testList_0/openjdkbinary/j2sdk-image/Contents/Home/jre/lib/default/libj9jit29.dylib [2021-06-17T08:37:19.574Z] LT stderr Module_base_address=000000000D6FE000 Symbol=..@109.done [2021-06-17T08:37:19.574Z] LT stderr Symbol_address=000000000D9BB4C0 [2021-06-17T08:37:19.574Z] LT stderr Target=2_90_20210616_155 (Mac OS X 10.14.6) [2021-06-17T08:37:19.574Z] LT stderr CPU=amd64 (12 logical CPUs) (0x400000000 RAM) [2021-06-17T08:37:19.574Z] LT stderr ----------- Stack Backtrace ----------- [2021-06-17T08:37:19.574Z] LT stderr --------------------------------------- [2021-06-17T08:37:19.574Z] LT stderr JVMDUMP039I Processing dump event "gpf", detail "" at 2021/06/17 04:37:18 - please wait. [2021-06-17T08:37:19.574Z] LT stderr JVMDUMP032I JVM requested System dump using '/Users/jenkins/workspace/Test_openjdk8_j9_special.system_x86-64_mac_testList_0/aqa-tests/TKG/output_16239023732122/MauveMultiThrdLoad_special_5m_22/20210617-043704-MauveMultiThrdLoadTrc/results/core.20210617.043718.95385.0001.dmp' in response to an event [2021-06-17T08:37:24.943Z] LT stderr JVMDUMP010I System dump written to /Users/jenkins/workspace/Test_openjdk8_j9_special.system_x86-64_mac_testList_0/aqa-tests/TKG/output_16239023732122/MauveMultiThrdLoad_special_5m_22/20210617-043704-MauveMultiThrdLoadTrc/results/core.20210617.043718.95385.0001.dmp [2021-06-17T08:37:24.943Z] LT stderr 0000000016823300: Object neither in heap nor stack-allocated in thread load-7 [2021-06-17T08:37:24.943Z] LT stderr 0000000016823300: O-Slot=00000000168592D0 [2021-06-17T08:37:24.943Z] LT stderr 0000000016823300: O-Slot value=00000000FFE39700 [2021-06-17T08:37:24.943Z] LT stderr 0000000016823300: PC=0000000000000005 [2021-06-17T08:37:24.943Z] LT stderr 0000000016823300: framesWalked=0 [2021-06-17T08:37:24.943Z] LT stderr 0000000016823300: arg0EA=00000000168592C8 [2021-06-17T08:37:24.943Z] LT stderr 0000000016823300: walkSP=00000000168592D0 [2021-06-17T08:37:24.943Z] LT stderr 0000000016823300: literals=0000000000000000 [2021-06-17T08:37:24.943Z] LT stderr 0000000016823300: jitInfo=0000000000000000 [2021-06-17T08:37:24.943Z] LT stderr 0000000016823300: stack=00000000168527F0-0000000016859850 [2021-06-17T08:37:24.943Z] STF 04:37:22.742 - Found dump at: /Users/jenkins/workspace/Test_openjdk8_j9_special.system_x86-64_mac_testList_0/aqa-tests/TKG/output_16239023732122/MauveMultiThrdLoad_special_5m_22/20210617-043704-MauveMultiThrdLoadTrc/results/core.20210617.043718.95385.0001.dmp [2021-06-17T08:37:24.943Z] STF 04:37:22.742 - **FAILED** Process LT ended with exit code (255) and not the expected exit code/s (0) [2021-06-17T08:37:24.943Z] LT stderr core file generated - /Users/jenkins/workspace/Test_openjdk8_j9_special.system_x86-64_mac_testList_0/aqa-tests/TKG/output_16239023732122/MauveMultiThrdLoad_special_5m_22/20210617-043704-MauveMultiThrdLoadTrc/results/core.20210617.043718.95385.0001.dmp [2021-06-17T08:37:24.943Z] STF 04:37:22.742 - Monitoring Report Summary: [2021-06-17T08:37:24.943Z] STF 04:37:22.742 - o Process LT has crashed unexpectedly [2021-06-17T08:37:24.943Z] STF 04:37:22.742 - Killing processes: LT [2021-06-17T08:37:24.943Z] STF 04:37:22.742 - o Process LT pid 95385 is not running [2021-06-17T08:37:24.943Z] **FAILED** at step 1 (Run Mauve load test). Expected return value=0 Actual=1 at /Users/jenkins/workspace/Test_openjdk8_j9_special.system_x86-64_mac_testList_0/aqa-tests/TKG/../TKG/output_16239023732122/MauveMultiThrdLoad_special_5m_22/20210617-043704-MauveMultiThrdLoadTrc/execute.pl line 97. [2021-06-17T08:37:24.943Z] STF 04:37:22.785 - **FAILED** execute script failed. Expected return value=0 Actual=1 [2021-06-17T08:37:24.943Z] STF 04:37:22.785 - [2021-06-17T08:37:24.943Z] STF 04:37:22.785 - ==================== T E A R D O W N ==================== [2021-06-17T08:37:24.943Z] STF 04:37:22.785 - Running teardown: perl /Users/jenkins/workspace/Test_openjdk8_j9_special.system_x86-64_mac_testList_0/aqa-tests/TKG/../TKG/output_16239023732122/MauveMultiThrdLoad_special_5m_22/20210617-043704-MauveMultiThrdLoadTrc/tearDown.pl [2021-06-17T08:37:24.943Z] STF 04:37:22.834 - TEARDOWN stage completed [2021-06-17T08:37:24.943Z] STF 04:37:22.837 - [2021-06-17T08:37:24.943Z] STF 04:37:22.837 - ===================== R E S U L T S ===================== [2021-06-17T08:37:24.943Z] STF 04:37:22.837 - Stage results: [2021-06-17T08:37:24.944Z] STF 04:37:22.837 - setUp: pass [2021-06-17T08:37:24.944Z] STF 04:37:22.837 - execute: *fail* [2021-06-17T08:37:24.944Z] STF 04:37:22.837 - teardown: pass [2021-06-17T08:37:24.944Z] STF 04:37:22.837 - [2021-06-17T08:37:24.944Z] STF 04:37:22.837 - Overall result: **FAILED** [2021-06-17T08:37:24.944Z] [2021-06-17T08:37:24.944Z] MauveMultiThrdLoad_special_5m_22_FAILED ``` Launched a 30x grinder at `Grinder/16361/` and a grinder on same machine at `Grinder/16362/`
1.0
JDK8 MacOS MauveMultiThrdLoad_special_5m_22_FAILED - Segmentation error vmState=0x00000000 Module=default/libj9jit29.dylib - Failure link ------------ From an internal build `Test_openjdk8_j9_special.system_x86-64_mac_testList_0/44/` (`osxrt11`): ``` openjdk version "1.8.0_302-beta" OpenJDK Runtime Environment (build 1.8.0_302-beta-202106170233-b05) Eclipse OpenJ9 VM (build master-0385f0737, JRE 1.8.0 Mac OS X amd64-64-Bit Compressed References 20210616_155 (JIT enabled, AOT enabled) OpenJ9 - 0385f0737 OMR - 07e7658a9 JCL - be819230 based on jdk8u302-b05) ``` [Rerun in Grinder](https://hyc-runtimes-jenkins.swg-devops.com/job/Grinder/parambuild/?SDK_RESOURCE=upstream&amp;TARGET=MauveMultiThrdLoad_special_5m_22&amp;TEST_FLAG=&amp;UPSTREAM_TEST_JOB_NAME=Test_openjdk8_j9_special.system_x86-64_mac&amp;DOCKER_REQUIRED=false&amp;ACTIVE_NODE_TIMEOUT=0&amp;VENDOR_TEST_DIRS=&amp;EXTRA_DOCKER_ARGS=&amp;TKG_OWNER_BRANCH=AdoptOpenJDK%3Amaster&amp;TEST_PARALLELIZATION_PARAMS=&amp;OPENJ9_SYSTEMTEST_OWNER_BRANCH=eclipse%3Amaster&amp;PLATFORM=x86-64_mac&amp;GENERATE_JOBS=true&amp;KEEP_REPORTDIR=false&amp;PERSONAL_BUILD=false&amp;ADOPTOPENJDK_REPO=https%3A%2F%2Fgithub.com%2Fadoptium%2Faqa-tests.git&amp;LABEL=&amp;TEST_OPTIONS_PARAMS=&amp;EXTRA_OPTIONS=&amp;CUSTOMIZED_SDK_URL=+https%3A%2F%2Fna.artifactory.swg-devops.com%2Fartifactory%2Fsys-rt-generic-local%2Fhyc-runtimes-jenkins.swg-devops.com%2Fbuild-scripts%2Fjobs%2Fjdk8u%2Fjdk8u-mac-x64-openj9%2F155%2FOpenJDK8U-jdk_x64_mac_openj9_2021-06-17-02-30.tar.gz+https%3A%2F%2Fna.artifactory.swg-devops.com%2Fartifactory%2Fsys-rt-generic-local%2Fhyc-runtimes-jenkins.swg-devops.com%2Fbuild-scripts%2Fjobs%2Fjdk8u%2Fjdk8u-mac-x64-openj9%2F155%2FOpenJDK8U-jre_x64_mac_openj9_2021-06-17-02-30.tar.gz+https%3A%2F%2Fna.artifactory.swg-devops.com%2Fartifactory%2Fsys-rt-generic-local%2Fhyc-runtimes-jenkins.swg-devops.com%2Fbuild-scripts%2Fjobs%2Fjdk8u%2Fjdk8u-mac-x64-openj9%2F155%2FOpenJDK8U-testimage_x64_mac_openj9_2021-06-17-02-30.tar.gz+https%3A%2F%2Fna.artifactory.swg-devops.com%2Fartifactory%2Fsys-rt-generic-local%2Fhyc-runtimes-jenkins.swg-devops.com%2Fbuild-scripts%2Fjobs%2Fjdk8u%2Fjdk8u-mac-x64-openj9%2F155%2FOpenJDK8U-debugimage_x64_mac_openj9_2021-06-17-02-30.tar.gz&amp;BUILD_IDENTIFIER=&amp;NON_AQA_TEST_REPOS_HELP_TEXT=&amp;ADOPTOPENJDK_BRANCH=master&amp;LIGHT_WEIGHT_CHECKOUT=true&amp;NON_AQA_TEST_REPOS=&amp;ARTIFACTORY_SERVER=na.artifactory.swg-devops&amp;TEST_REPO_PARAMS=&amp;TEST_SELECTION_PARAMS=&amp;TEST_PARALLELIZATION_PARAMS_HELP_TEXT=&amp;JDK_SELECTION_PARAMS=&amp;KEEP_WORKSPACE=false&amp;USER_CREDENTIALS_ID=&amp;JDK_VERSION=8&amp;ITERATIONS=1&amp;VENDOR_TEST_REPOS=&amp;JDK_REPO=https%3A%2F%2Fgithub.com%2Fibmruntimes%2Fopenj9-openjdk-jdk8&amp;PLATFORM_AND_MACHINE_HELP_TEXT=&amp;RELEASE_TAG=&amp;OPENJ9_BRANCH=master&amp;OPENJ9_SHA=&amp;JCK_GIT_REPO=&amp;VENDOR_TEST_BRANCHES=&amp;OPENJ9_REPO=https%3A%2F%2Fgithub.com%2Feclipse-openj9%2Fopenj9.git&amp;UPSTREAM_JOB_NAME=&amp;CLOUD_PROVIDER=&amp;PLATFORM_AND_MACHINE=&amp;CUSTOM_TARGET=&amp;VENDOR_TEST_SHAS=&amp;JDK_BRANCH=openj9&amp;LABEL_ADDITION=&amp;ARTIFACTORY_REPO=&amp;ARTIFACTORY_ROOT_DIR=&amp;POST_RUN_PARAMS_HELP_TEXT=&amp;UPSTREAM_TEST_JOB_NUMBER=45&amp;DOCKERIMAGE_TAG=&amp;JDK_SELECTION_PARAMS_HELP_TEXT=&amp;JDK_IMPL=openj9&amp;TEST_TIME=&amp;SSH_AGENT_CREDENTIAL=83181e25-eea4-4f55-8b3e-e79615733226&amp;AUTO_DETECT=true&amp;TKG_SHA=&amp;TEST_SELECTION_PARAMS_HELP_TEXT=&amp;ADOPTOPENJDK_SYSTEMTEST_OWNER_BRANCH=adoptium%3Amaster&amp;CUSTOMIZED_SDK_URL_CREDENTIAL_ID=4e18ffe7-b1b1-4272-9979-99769b68bcc2&amp;OPENJDK_SHA=&amp;NUM_MACHINES=&amp;TRSS_URL=http%3A%2F%2Ftrss1.fyre.ibm.com&amp;BUILD_LIST=system&amp;UPSTREAM_JOB_NUMBER=&amp;STF_OWNER_BRANCH=adoptium%3Amaster&amp;TEST_REPO_PARAMS_HELP_TEXT=&amp;POST_RUN_PARAMS=&amp;TIME_LIMIT=20&amp;TEST_OPTIONS_PARAMS_HELP_TEXT=&amp;JVM_OPTIONS=&amp;PARALLEL=None) Optional info ------------- Failure output (captured from console output) --------------------------------------------- ``` [2021-06-17T08:37:05.525Z] Running test MauveMultiThrdLoad_special_5m_22 ... [2021-06-17T08:37:05.525Z] =============================================== [2021-06-17T08:37:05.525Z] MauveMultiThrdLoad_special_5m_22 Start Time: Thu Jun 17 04:37:04 2021 Epoch Time (ms): 1623919024447 [2021-06-17T08:37:05.525Z] "/Users/jenkins/workspace/Test_openjdk8_j9_special.system_x86-64_mac_testList_0/openjdkbinary/j2sdk-image/Contents/Home/bin/..//bin/java" -Xshareclasses:destroyAll; "/Users/jenkins/workspace/Test_openjdk8_j9_special.system_x86-64_mac_testList_0/openjdkbinary/j2sdk-image/Contents/Home/bin/..//bin/java" -Xshareclasses:groupAccess,destroyAll; echo "cache cleanup done"; [2021-06-17T08:37:05.525Z] JVMSHRC005I No shared class caches available [2021-06-17T08:37:05.525Z] JVMSHRC005I No shared class caches available [2021-06-17T08:37:05.525Z] cache cleanup done [2021-06-17T08:37:05.525Z] variation: Mode688 [2021-06-17T08:37:05.525Z] JVM_OPTIONS: -Xcompressedrefs -Xjit:count=0 -Xgcpolicy:gencon -Xaggressive -Xconcurrentlevel0 [2021-06-17T08:37:19.574Z] LT stderr Unhandled exception [2021-06-17T08:37:19.574Z] LT stderr Type=Segmentation error vmState=0x00000000 [2021-06-17T08:37:19.574Z] LT stderr J9Generic_Signal_Number=00000018 Signal_Number=0000000b Error_Value=00000000 Signal_Code=00000001 [2021-06-17T08:37:19.574Z] LT stderr Handler1=000000000D236CE0 Handler2=000000000D510C10 InaccessibleAddress=00000000FFE398B8 [2021-06-17T08:37:19.574Z] LT stderr RDI=000000002351A2DB RSI=000000002351A2E0 RAX=0000000000000048 RBX=0000000016883300 [2021-06-17T08:37:19.574Z] LT stderr RCX=00000000FFE39700 RDX=000000002351A2E0 R8=0000000016883638 R9=00000007FFE39410 [2021-06-17T08:37:19.574Z] LT stderr R10=0000000700099F80 R11=000000000B9EE050 R12=00000007FFCB1760 R13=00000007FFCB2D88 [2021-06-17T08:37:19.574Z] LT stderr R14=000000000000000A R15=00000007FFCB2D30 [2021-06-17T08:37:19.574Z] LT stderr RIP=000000000D9BB4D0 GS=0000 FS=0000 RSP=00000000168592D0 [2021-06-17T08:37:19.574Z] LT stderr RFlags=0000000000010217 CS=002B RBP=0000000016823300 ERR=FFE398B800000004 [2021-06-17T08:37:19.574Z] LT stderr TRAPNO=000000040000000E CPU=98B8000000040000 FAULTVADDR=00000000FFE398B8 [2021-06-17T08:37:19.574Z] LT stderr XMM0 0000000000000002 (f: 2.000000, d: 9.881313e-324) [2021-06-17T08:37:19.574Z] LT stderr XMM1 0000000016859318 (f: 377852704.000000, d: 1.866840e-315) [2021-06-17T08:37:19.574Z] LT stderr XMM2 0000000000000000 (f: 0.000000, d: 0.000000e+00) [2021-06-17T08:37:19.574Z] LT stderr XMM3 0000000000000000 (f: 0.000000, d: 0.000000e+00) [2021-06-17T08:37:19.574Z] LT stderr XMM4 0000000000000000 (f: 0.000000, d: 0.000000e+00) [2021-06-17T08:37:19.574Z] LT stderr XMM5 00000000001c0000 (f: 1835008.000000, d: 9.066144e-318) [2021-06-17T08:37:19.574Z] LT stderr XMM6 3fd89a3386c1425b (f: 2260812288.000000, d: 3.844117e-01) [2021-06-17T08:37:19.574Z] LT stderr XMM7 401e7f9c1e980fa9 (f: 513281952.000000, d: 7.624619e+00) [2021-06-17T08:37:19.574Z] LT stderr XMM8 ae5ff75e430f4000 (f: 1125072896.000000, d: -2.571067e-85) [2021-06-17T08:37:19.574Z] LT stderr XMM9 0000000000000000 (f: 0.000000, d: 0.000000e+00) [2021-06-17T08:37:19.574Z] LT stderr XMM10 0000000000000006 (f: 6.000000, d: 2.964394e-323) [2021-06-17T08:37:19.574Z] LT stderr XMM11 0000000000000008 (f: 8.000000, d: 3.952525e-323) [2021-06-17T08:37:19.574Z] LT stderr XMM12 0000000000000000 (f: 0.000000, d: 0.000000e+00) [2021-06-17T08:37:19.574Z] LT stderr XMM13 0000000000000000 (f: 0.000000, d: 0.000000e+00) [2021-06-17T08:37:19.574Z] LT stderr XMM14 0000000000000000 (f: 0.000000, d: 0.000000e+00) [2021-06-17T08:37:19.574Z] LT stderr XMM15 0000000000000000 (f: 0.000000, d: 0.000000e+00) [2021-06-17T08:37:19.574Z] LT stderr Module=/Users/jenkins/workspace/Test_openjdk8_j9_special.system_x86-64_mac_testList_0/openjdkbinary/j2sdk-image/Contents/Home/jre/lib/default/libj9jit29.dylib [2021-06-17T08:37:19.574Z] LT stderr Module_base_address=000000000D6FE000 Symbol=..@109.done [2021-06-17T08:37:19.574Z] LT stderr Symbol_address=000000000D9BB4C0 [2021-06-17T08:37:19.574Z] LT stderr Target=2_90_20210616_155 (Mac OS X 10.14.6) [2021-06-17T08:37:19.574Z] LT stderr CPU=amd64 (12 logical CPUs) (0x400000000 RAM) [2021-06-17T08:37:19.574Z] LT stderr ----------- Stack Backtrace ----------- [2021-06-17T08:37:19.574Z] LT stderr --------------------------------------- [2021-06-17T08:37:19.574Z] LT stderr JVMDUMP039I Processing dump event "gpf", detail "" at 2021/06/17 04:37:18 - please wait. [2021-06-17T08:37:19.574Z] LT stderr JVMDUMP032I JVM requested System dump using '/Users/jenkins/workspace/Test_openjdk8_j9_special.system_x86-64_mac_testList_0/aqa-tests/TKG/output_16239023732122/MauveMultiThrdLoad_special_5m_22/20210617-043704-MauveMultiThrdLoadTrc/results/core.20210617.043718.95385.0001.dmp' in response to an event [2021-06-17T08:37:24.943Z] LT stderr JVMDUMP010I System dump written to /Users/jenkins/workspace/Test_openjdk8_j9_special.system_x86-64_mac_testList_0/aqa-tests/TKG/output_16239023732122/MauveMultiThrdLoad_special_5m_22/20210617-043704-MauveMultiThrdLoadTrc/results/core.20210617.043718.95385.0001.dmp [2021-06-17T08:37:24.943Z] LT stderr 0000000016823300: Object neither in heap nor stack-allocated in thread load-7 [2021-06-17T08:37:24.943Z] LT stderr 0000000016823300: O-Slot=00000000168592D0 [2021-06-17T08:37:24.943Z] LT stderr 0000000016823300: O-Slot value=00000000FFE39700 [2021-06-17T08:37:24.943Z] LT stderr 0000000016823300: PC=0000000000000005 [2021-06-17T08:37:24.943Z] LT stderr 0000000016823300: framesWalked=0 [2021-06-17T08:37:24.943Z] LT stderr 0000000016823300: arg0EA=00000000168592C8 [2021-06-17T08:37:24.943Z] LT stderr 0000000016823300: walkSP=00000000168592D0 [2021-06-17T08:37:24.943Z] LT stderr 0000000016823300: literals=0000000000000000 [2021-06-17T08:37:24.943Z] LT stderr 0000000016823300: jitInfo=0000000000000000 [2021-06-17T08:37:24.943Z] LT stderr 0000000016823300: stack=00000000168527F0-0000000016859850 [2021-06-17T08:37:24.943Z] STF 04:37:22.742 - Found dump at: /Users/jenkins/workspace/Test_openjdk8_j9_special.system_x86-64_mac_testList_0/aqa-tests/TKG/output_16239023732122/MauveMultiThrdLoad_special_5m_22/20210617-043704-MauveMultiThrdLoadTrc/results/core.20210617.043718.95385.0001.dmp [2021-06-17T08:37:24.943Z] STF 04:37:22.742 - **FAILED** Process LT ended with exit code (255) and not the expected exit code/s (0) [2021-06-17T08:37:24.943Z] LT stderr core file generated - /Users/jenkins/workspace/Test_openjdk8_j9_special.system_x86-64_mac_testList_0/aqa-tests/TKG/output_16239023732122/MauveMultiThrdLoad_special_5m_22/20210617-043704-MauveMultiThrdLoadTrc/results/core.20210617.043718.95385.0001.dmp [2021-06-17T08:37:24.943Z] STF 04:37:22.742 - Monitoring Report Summary: [2021-06-17T08:37:24.943Z] STF 04:37:22.742 - o Process LT has crashed unexpectedly [2021-06-17T08:37:24.943Z] STF 04:37:22.742 - Killing processes: LT [2021-06-17T08:37:24.943Z] STF 04:37:22.742 - o Process LT pid 95385 is not running [2021-06-17T08:37:24.943Z] **FAILED** at step 1 (Run Mauve load test). Expected return value=0 Actual=1 at /Users/jenkins/workspace/Test_openjdk8_j9_special.system_x86-64_mac_testList_0/aqa-tests/TKG/../TKG/output_16239023732122/MauveMultiThrdLoad_special_5m_22/20210617-043704-MauveMultiThrdLoadTrc/execute.pl line 97. [2021-06-17T08:37:24.943Z] STF 04:37:22.785 - **FAILED** execute script failed. Expected return value=0 Actual=1 [2021-06-17T08:37:24.943Z] STF 04:37:22.785 - [2021-06-17T08:37:24.943Z] STF 04:37:22.785 - ==================== T E A R D O W N ==================== [2021-06-17T08:37:24.943Z] STF 04:37:22.785 - Running teardown: perl /Users/jenkins/workspace/Test_openjdk8_j9_special.system_x86-64_mac_testList_0/aqa-tests/TKG/../TKG/output_16239023732122/MauveMultiThrdLoad_special_5m_22/20210617-043704-MauveMultiThrdLoadTrc/tearDown.pl [2021-06-17T08:37:24.943Z] STF 04:37:22.834 - TEARDOWN stage completed [2021-06-17T08:37:24.943Z] STF 04:37:22.837 - [2021-06-17T08:37:24.943Z] STF 04:37:22.837 - ===================== R E S U L T S ===================== [2021-06-17T08:37:24.943Z] STF 04:37:22.837 - Stage results: [2021-06-17T08:37:24.944Z] STF 04:37:22.837 - setUp: pass [2021-06-17T08:37:24.944Z] STF 04:37:22.837 - execute: *fail* [2021-06-17T08:37:24.944Z] STF 04:37:22.837 - teardown: pass [2021-06-17T08:37:24.944Z] STF 04:37:22.837 - [2021-06-17T08:37:24.944Z] STF 04:37:22.837 - Overall result: **FAILED** [2021-06-17T08:37:24.944Z] [2021-06-17T08:37:24.944Z] MauveMultiThrdLoad_special_5m_22_FAILED ``` Launched a 30x grinder at `Grinder/16361/` and a grinder on same machine at `Grinder/16362/`
non_process
macos mauvemultithrdload special failed segmentation error vmstate module default dylib failure link from an internal build test special system mac testlist openjdk version beta openjdk runtime environment build beta eclipse vm build master jre mac os x bit compressed references jit enabled aot enabled omr jcl based on optional info failure output captured from console output running test mauvemultithrdload special mauvemultithrdload special start time thu jun epoch time ms users jenkins workspace test special system mac testlist openjdkbinary image contents home bin bin java xshareclasses destroyall users jenkins workspace test special system mac testlist openjdkbinary image contents home bin bin java xshareclasses groupaccess destroyall echo cache cleanup done no shared class caches available no shared class caches available cache cleanup done variation jvm options xcompressedrefs xjit count xgcpolicy gencon xaggressive lt stderr unhandled exception lt stderr type segmentation error vmstate lt stderr signal number signal number error value signal code lt stderr inaccessibleaddress lt stderr rdi rsi rax rbx lt stderr rcx rdx lt stderr lt stderr lt stderr rip gs fs rsp lt stderr rflags cs rbp err lt stderr trapno cpu faultvaddr lt stderr f d lt stderr f d lt stderr f d lt stderr f d lt stderr f d lt stderr f d lt stderr f d lt stderr f d lt stderr f d lt stderr f d lt stderr f d lt stderr f d lt stderr f d lt stderr f d lt stderr f d lt stderr f d lt stderr module users jenkins workspace test special system mac testlist openjdkbinary image contents home jre lib default dylib lt stderr module base address symbol done lt stderr symbol address lt stderr target mac os x lt stderr cpu logical cpus ram lt stderr stack backtrace lt stderr lt stderr processing dump event gpf detail at please wait lt stderr jvm requested system dump using users jenkins workspace test special system mac testlist aqa tests tkg output mauvemultithrdload special mauvemultithrdloadtrc results core dmp in response to an event lt stderr system dump written to users jenkins workspace test special system mac testlist aqa tests tkg output mauvemultithrdload special mauvemultithrdloadtrc results core dmp lt stderr object neither in heap nor stack allocated in thread load lt stderr o slot lt stderr o slot value lt stderr pc lt stderr frameswalked lt stderr lt stderr walksp lt stderr literals lt stderr jitinfo lt stderr stack stf found dump at users jenkins workspace test special system mac testlist aqa tests tkg output mauvemultithrdload special mauvemultithrdloadtrc results core dmp stf failed process lt ended with exit code and not the expected exit code s lt stderr core file generated users jenkins workspace test special system mac testlist aqa tests tkg output mauvemultithrdload special mauvemultithrdloadtrc results core dmp stf monitoring report summary stf o process lt has crashed unexpectedly stf killing processes lt stf o process lt pid is not running failed at step run mauve load test expected return value actual at users jenkins workspace test special system mac testlist aqa tests tkg tkg output mauvemultithrdload special mauvemultithrdloadtrc execute pl line stf failed execute script failed expected return value actual stf stf t e a r d o w n stf running teardown perl users jenkins workspace test special system mac testlist aqa tests tkg tkg output mauvemultithrdload special mauvemultithrdloadtrc teardown pl stf teardown stage completed stf stf r e s u l t s stf stage results stf setup pass stf execute fail stf teardown pass stf stf overall result failed mauvemultithrdload special failed launched a grinder at grinder and a grinder on same machine at grinder
0
77,027
15,496,255,887
IssuesEvent
2021-03-11 02:20:33
n-devs/BuySellCar-Online
https://api.github.com/repos/n-devs/BuySellCar-Online
opened
WS-2019-0424 (Medium) detected in elliptic-6.5.0.tgz
security vulnerability
## WS-2019-0424 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>elliptic-6.5.0.tgz</b></p></summary> <p>EC cryptography</p> <p>Library home page: <a href="https://registry.npmjs.org/elliptic/-/elliptic-6.5.0.tgz">https://registry.npmjs.org/elliptic/-/elliptic-6.5.0.tgz</a></p> <p>Path to dependency file: /BuySellCar-Online/package.json</p> <p>Path to vulnerable library: BuySellCar-Online/node_modules/elliptic/package.json</p> <p> Dependency Hierarchy: - react-scripts-2.1.8.tgz (Root Library) - webpack-4.28.3.tgz - node-libs-browser-2.2.1.tgz - crypto-browserify-3.12.0.tgz - browserify-sign-4.0.4.tgz - :x: **elliptic-6.5.0.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> all versions of elliptic are vulnerable to Timing Attack through side-channels. <p>Publish Date: 2019-11-13 <p>URL: <a href=https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a>WS-2019-0424</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Adjacent - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2019-0424 (Medium) detected in elliptic-6.5.0.tgz - ## WS-2019-0424 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>elliptic-6.5.0.tgz</b></p></summary> <p>EC cryptography</p> <p>Library home page: <a href="https://registry.npmjs.org/elliptic/-/elliptic-6.5.0.tgz">https://registry.npmjs.org/elliptic/-/elliptic-6.5.0.tgz</a></p> <p>Path to dependency file: /BuySellCar-Online/package.json</p> <p>Path to vulnerable library: BuySellCar-Online/node_modules/elliptic/package.json</p> <p> Dependency Hierarchy: - react-scripts-2.1.8.tgz (Root Library) - webpack-4.28.3.tgz - node-libs-browser-2.2.1.tgz - crypto-browserify-3.12.0.tgz - browserify-sign-4.0.4.tgz - :x: **elliptic-6.5.0.tgz** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> all versions of elliptic are vulnerable to Timing Attack through side-channels. <p>Publish Date: 2019-11-13 <p>URL: <a href=https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a>WS-2019-0424</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Adjacent - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
ws medium detected in elliptic tgz ws medium severity vulnerability vulnerable library elliptic tgz ec cryptography library home page a href path to dependency file buysellcar online package json path to vulnerable library buysellcar online node modules elliptic package json dependency hierarchy react scripts tgz root library webpack tgz node libs browser tgz crypto browserify tgz browserify sign tgz x elliptic tgz vulnerable library vulnerability details all versions of elliptic are vulnerable to timing attack through side channels publish date url a href cvss score details base score metrics exploitability metrics attack vector adjacent attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact high availability impact none for more information on scores click a href step up your open source security game with whitesource
0
4,551
7,381,896,768
IssuesEvent
2018-03-15 01:27:30
mlmasters/toxic-comments
https://api.github.com/repos/mlmasters/toxic-comments
closed
Identify and handle outliers
preprocessing
There are outliers in the data which may hinder our efforts at predicting toxicity. For example, there are some extremely long comments that may skew word counts. One complication of this is the fact that there is a strong relationship between toxicity and being an outlier on comment length (i.e., outliers on comment length tend to be toxic). Questions: - Does using TF-IDF or binary t-d matrix improve the outlier situation? - How should we handle outliers on length given the above? - Are there outliers on other variables? If so, what should we do about them?
1.0
Identify and handle outliers - There are outliers in the data which may hinder our efforts at predicting toxicity. For example, there are some extremely long comments that may skew word counts. One complication of this is the fact that there is a strong relationship between toxicity and being an outlier on comment length (i.e., outliers on comment length tend to be toxic). Questions: - Does using TF-IDF or binary t-d matrix improve the outlier situation? - How should we handle outliers on length given the above? - Are there outliers on other variables? If so, what should we do about them?
process
identify and handle outliers there are outliers in the data which may hinder our efforts at predicting toxicity for example there are some extremely long comments that may skew word counts one complication of this is the fact that there is a strong relationship between toxicity and being an outlier on comment length i e outliers on comment length tend to be toxic questions does using tf idf or binary t d matrix improve the outlier situation how should we handle outliers on length given the above are there outliers on other variables if so what should we do about them
1
21,577
29,934,153,039
IssuesEvent
2023-06-22 11:36:21
NationalSecurityAgency/ghidra
https://api.github.com/repos/NationalSecurityAgency/ghidra
closed
ARM vdup.32 is decoded incorrectly
Feature: Processor/ARM Status: Internal
**Describe the bug** The ARM `vdup.32` instruction - more specifically, the source register - is decoded incorrectly. For example, `80 EE 10 0B` (`vdup.32 d0, r0`) will be decoded as `vdup.32 d0, lr`, leading to nonsensical disassembly. The built-in assembler (accessed via `Patch Instruction`) will also refuse to assemble any variant of the instruction that doesn't have `lr` as the source register. **To Reproduce** Steps to reproduce the behavior: 1. Import an ARM binary exhibiting the faulty instruction (tested with `ARM:v7:32:little:default` and `ARM:v8T:32:little:default`) 3. Disassemble the instructions 4. See error **Expected behavior** The source register is decoded properly and the disassembly is correct. **Screenshots** A sample program containing different variants of the instruction. While the destination operand is decoded properly, the source operand is always decoded as `lr` regardless of the instruction's content. ![Sample program](https://github.com/NationalSecurityAgency/ghidra/assets/32939301/bbf8aabc-c255-48a8-a692-7f2744b3d7d9) A real-world application exhibiting this problem. While a floating point value should be shown stored in the disassembly, `in_lr` is displayed instead due to this bug. ![RC2 Camera_InitFovSystem](https://github.com/NationalSecurityAgency/ghidra/assets/32939301/41547e69-b717-4854-9a48-a85dc4042ea7) The correct disassembly, performed by Capstone (made via http://shell-storm.org/online/Online-Assembler-and-Disassembler/) ![Correct disassembly from Capstone](https://github.com/NationalSecurityAgency/ghidra/assets/32939301/5d9e1ace-5f90-4e5a-ae80-fe151392d9d8) **Attachments** A sample program (shown in the first screenshot from previous section) containing a few variants of the instruction. [bad_vdup32.zip](https://github.com/NationalSecurityAgency/ghidra/files/11657833/bad_vdup32.zip) **Environment (please complete the following information):** - OS: Windows 10 Pro 22H2 build 19045.2965 - Java Version: 17.0.7 - Ghidra Version: 10.3 - Ghidra Origin: Official GitHub distro
1.0
ARM vdup.32 is decoded incorrectly - **Describe the bug** The ARM `vdup.32` instruction - more specifically, the source register - is decoded incorrectly. For example, `80 EE 10 0B` (`vdup.32 d0, r0`) will be decoded as `vdup.32 d0, lr`, leading to nonsensical disassembly. The built-in assembler (accessed via `Patch Instruction`) will also refuse to assemble any variant of the instruction that doesn't have `lr` as the source register. **To Reproduce** Steps to reproduce the behavior: 1. Import an ARM binary exhibiting the faulty instruction (tested with `ARM:v7:32:little:default` and `ARM:v8T:32:little:default`) 3. Disassemble the instructions 4. See error **Expected behavior** The source register is decoded properly and the disassembly is correct. **Screenshots** A sample program containing different variants of the instruction. While the destination operand is decoded properly, the source operand is always decoded as `lr` regardless of the instruction's content. ![Sample program](https://github.com/NationalSecurityAgency/ghidra/assets/32939301/bbf8aabc-c255-48a8-a692-7f2744b3d7d9) A real-world application exhibiting this problem. While a floating point value should be shown stored in the disassembly, `in_lr` is displayed instead due to this bug. ![RC2 Camera_InitFovSystem](https://github.com/NationalSecurityAgency/ghidra/assets/32939301/41547e69-b717-4854-9a48-a85dc4042ea7) The correct disassembly, performed by Capstone (made via http://shell-storm.org/online/Online-Assembler-and-Disassembler/) ![Correct disassembly from Capstone](https://github.com/NationalSecurityAgency/ghidra/assets/32939301/5d9e1ace-5f90-4e5a-ae80-fe151392d9d8) **Attachments** A sample program (shown in the first screenshot from previous section) containing a few variants of the instruction. [bad_vdup32.zip](https://github.com/NationalSecurityAgency/ghidra/files/11657833/bad_vdup32.zip) **Environment (please complete the following information):** - OS: Windows 10 Pro 22H2 build 19045.2965 - Java Version: 17.0.7 - Ghidra Version: 10.3 - Ghidra Origin: Official GitHub distro
process
arm vdup is decoded incorrectly describe the bug the arm vdup instruction more specifically the source register is decoded incorrectly for example ee vdup will be decoded as vdup lr leading to nonsensical disassembly the built in assembler accessed via patch instruction will also refuse to assemble any variant of the instruction that doesn t have lr as the source register to reproduce steps to reproduce the behavior import an arm binary exhibiting the faulty instruction tested with arm little default and arm little default disassemble the instructions see error expected behavior the source register is decoded properly and the disassembly is correct screenshots a sample program containing different variants of the instruction while the destination operand is decoded properly the source operand is always decoded as lr regardless of the instruction s content a real world application exhibiting this problem while a floating point value should be shown stored in the disassembly in lr is displayed instead due to this bug the correct disassembly performed by capstone made via attachments a sample program shown in the first screenshot from previous section containing a few variants of the instruction environment please complete the following information os windows pro build java version ghidra version ghidra origin official github distro
1
19,222
25,358,608,307
IssuesEvent
2022-11-20 16:33:22
streamnative/pulsar-spark
https://api.github.com/repos/streamnative/pulsar-spark
closed
[BUG]write data into pulsar, Literal must have a corresponding value to string, but class String found
type/bug compute/data-processing
**Describe the bug** when write data in pulsar,,use the follow code: ```scala def main(args: Array[String]): Unit = { val sparkSession = SparkSession.builder().appName("test-pulsar").master("local").getOrCreate() val startingOffsets = topicOffsets(Map("persistent://public/default/my-topic" -> MessageId.fromByteArray(Array(8,33,16,8)))) import sparkSession.implicits._ val ds = sparkSession.createDataset(1 to 10) .write .format("pulsar") .option("service.url", "pulsar://localhost:6650") .option("admin.url", "http://localhost:8080") .option("topic", "persistent://public/default/my-topic-2") .save() sparkSession.stop() } ``` ------ then, get a error ``` Exception in thread "main" java.lang.IllegalArgumentException: requirement failed: Literal must have a corresponding value to string, but class String found. at scala.Predef$.require(Predef.scala:277) at org.apache.spark.sql.catalyst.expressions.Literal$.validateLiteralValue(literals.scala:219) at org.apache.spark.sql.catalyst.expressions.Literal.<init>(literals.scala:296) at org.apache.spark.sql.pulsar.PulsarSinks$.$anonfun$validateQuery$2(PulsarSinks.scala:89) at scala.Option.getOrElse(Option.scala:121) at org.apache.spark.sql.pulsar.PulsarSinks$.validateQuery(PulsarSinks.scala:83) at org.apache.spark.sql.pulsar.PulsarProvider.createRelation(PulsarProvider.scala:185) at org.apache.spark.sql.execution.datasources.SaveIntoDataSourceCommand.run(SaveIntoDataSourceCommand.scala:46) at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:70) at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:68) at org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute(commands.scala:90) at org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1(SparkPlan.scala:180) at org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1(SparkPlan.scala:218) at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:215) at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:176) at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:132) at org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:131) at org.apache.spark.sql.DataFrameWriter.$anonfun$runCommand$1(DataFrameWriter.scala:989) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:103) at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:163) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:90) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:775) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64) at org.apache.spark.sql.DataFrameWriter.runCommand(DataFrameWriter.scala:989) at org.apache.spark.sql.DataFrameWriter.saveToV1Source(DataFrameWriter.scala:438) at org.apache.spark.sql.DataFrameWriter.saveInternal(DataFrameWriter.scala:415) at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:301) at pulsar.exProduce$.main(exProduce.scala:26) at pulsar.exProduce.main(exProduce.scala) 21/08/13 15:41:12 INFO SparkContext: Invoking stop() from shutdown hook 21/08/13 15:41:12 INFO SparkUI: Stopped Spark web UI at http://172.18.21.74:4040 21/08/13 15:41:12 INFO MapOutputTrackerMasterEndpoint: MapOutputTrackerMasterEndpoint stopped! ``` ------------------------- **To Reproduce** data must write into pulsar, and in topic can read the write data **Expected behavior**
1.0
[BUG]write data into pulsar, Literal must have a corresponding value to string, but class String found - **Describe the bug** when write data in pulsar,,use the follow code: ```scala def main(args: Array[String]): Unit = { val sparkSession = SparkSession.builder().appName("test-pulsar").master("local").getOrCreate() val startingOffsets = topicOffsets(Map("persistent://public/default/my-topic" -> MessageId.fromByteArray(Array(8,33,16,8)))) import sparkSession.implicits._ val ds = sparkSession.createDataset(1 to 10) .write .format("pulsar") .option("service.url", "pulsar://localhost:6650") .option("admin.url", "http://localhost:8080") .option("topic", "persistent://public/default/my-topic-2") .save() sparkSession.stop() } ``` ------ then, get a error ``` Exception in thread "main" java.lang.IllegalArgumentException: requirement failed: Literal must have a corresponding value to string, but class String found. at scala.Predef$.require(Predef.scala:277) at org.apache.spark.sql.catalyst.expressions.Literal$.validateLiteralValue(literals.scala:219) at org.apache.spark.sql.catalyst.expressions.Literal.<init>(literals.scala:296) at org.apache.spark.sql.pulsar.PulsarSinks$.$anonfun$validateQuery$2(PulsarSinks.scala:89) at scala.Option.getOrElse(Option.scala:121) at org.apache.spark.sql.pulsar.PulsarSinks$.validateQuery(PulsarSinks.scala:83) at org.apache.spark.sql.pulsar.PulsarProvider.createRelation(PulsarProvider.scala:185) at org.apache.spark.sql.execution.datasources.SaveIntoDataSourceCommand.run(SaveIntoDataSourceCommand.scala:46) at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:70) at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:68) at org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute(commands.scala:90) at org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1(SparkPlan.scala:180) at org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1(SparkPlan.scala:218) at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151) at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:215) at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:176) at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:132) at org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:131) at org.apache.spark.sql.DataFrameWriter.$anonfun$runCommand$1(DataFrameWriter.scala:989) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:103) at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:163) at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:90) at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:775) at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64) at org.apache.spark.sql.DataFrameWriter.runCommand(DataFrameWriter.scala:989) at org.apache.spark.sql.DataFrameWriter.saveToV1Source(DataFrameWriter.scala:438) at org.apache.spark.sql.DataFrameWriter.saveInternal(DataFrameWriter.scala:415) at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:301) at pulsar.exProduce$.main(exProduce.scala:26) at pulsar.exProduce.main(exProduce.scala) 21/08/13 15:41:12 INFO SparkContext: Invoking stop() from shutdown hook 21/08/13 15:41:12 INFO SparkUI: Stopped Spark web UI at http://172.18.21.74:4040 21/08/13 15:41:12 INFO MapOutputTrackerMasterEndpoint: MapOutputTrackerMasterEndpoint stopped! ``` ------------------------- **To Reproduce** data must write into pulsar, and in topic can read the write data **Expected behavior**
process
write data into pulsar literal must have a corresponding value to string but class string found describe the bug when write data in pulsar use the follow code scala def main args array unit val sparksession sparksession builder appname test pulsar master local getorcreate val startingoffsets topicoffsets map persistent public default my topic messageid frombytearray array import sparksession implicits val ds sparksession createdataset to write format pulsar option service url pulsar localhost option admin url option topic persistent public default my topic save sparksession stop then get a error exception in thread main java lang illegalargumentexception requirement failed literal must have a corresponding value to string but class string found at scala predef require predef scala at org apache spark sql catalyst expressions literal validateliteralvalue literals scala at org apache spark sql catalyst expressions literal literals scala at org apache spark sql pulsar pulsarsinks anonfun validatequery pulsarsinks scala at scala option getorelse option scala at org apache spark sql pulsar pulsarsinks validatequery pulsarsinks scala at org apache spark sql pulsar pulsarprovider createrelation pulsarprovider scala at org apache spark sql execution datasources saveintodatasourcecommand run saveintodatasourcecommand scala at org apache spark sql execution command executedcommandexec sideeffectresult lzycompute commands scala at org apache spark sql execution command executedcommandexec sideeffectresult commands scala at org apache spark sql execution command executedcommandexec doexecute commands scala at org apache spark sql execution sparkplan anonfun execute sparkplan scala at org apache spark sql execution sparkplan anonfun executequery sparkplan scala at org apache spark rdd rddoperationscope withscope rddoperationscope scala at org apache spark sql execution sparkplan executequery sparkplan scala at org apache spark sql execution sparkplan execute sparkplan scala at org apache spark sql execution queryexecution tordd lzycompute queryexecution scala at org apache spark sql execution queryexecution tordd queryexecution scala at org apache spark sql dataframewriter anonfun runcommand dataframewriter scala at org apache spark sql execution sqlexecution anonfun withnewexecutionid sqlexecution scala at org apache spark sql execution sqlexecution withsqlconfpropagated sqlexecution scala at org apache spark sql execution sqlexecution anonfun withnewexecutionid sqlexecution scala at org apache spark sql sparksession withactive sparksession scala at org apache spark sql execution sqlexecution withnewexecutionid sqlexecution scala at org apache spark sql dataframewriter runcommand dataframewriter scala at org apache spark sql dataframewriter dataframewriter scala at org apache spark sql dataframewriter saveinternal dataframewriter scala at org apache spark sql dataframewriter save dataframewriter scala at pulsar exproduce main exproduce scala at pulsar exproduce main exproduce scala info sparkcontext invoking stop from shutdown hook info sparkui stopped spark web ui at info mapoutputtrackermasterendpoint mapoutputtrackermasterendpoint stopped to reproduce data must write into pulsar and in topic can read the write data expected behavior
1
20,415
4,543,225,607
IssuesEvent
2016-09-10 01:34:41
brian-team/brian2
https://api.github.com/repos/brian-team/brian2
opened
Third tutorial
component: documentation enhancement
Would be good to have some 'advanced' topics like: * Subgroups * Linked variables * run_regularly * network_operation
1.0
Third tutorial - Would be good to have some 'advanced' topics like: * Subgroups * Linked variables * run_regularly * network_operation
non_process
third tutorial would be good to have some advanced topics like subgroups linked variables run regularly network operation
0
22,498
31,475,628,424
IssuesEvent
2023-08-30 10:28:35
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
closed
Duplicate error notification from terminal process non-zero termination
macos under-discussion terminal-process
<details><summary>Details</summary> <p> ``` Version: 1.80.0-insider Commit: 73eb619afb87cab4aa83a569259c891b6983c11e Date: 2023-06-19T05:56:57.953Z Electron: 22.3.11 ElectronBuildId: 21658788 Chromium: 108.0.5359.215 Node.js: 16.17.1 V8: 10.8.168.25-electron.0 OS: Darwin arm64 22.5.0 ``` </p> </details> Every now and then, exiting zsh by typing `exit` and hitting <kbd>Enter</kbd> results in an error notification. What just happened to me was that the same error notification came up in the window I did that, as well as another workspace window. I even had more workspace windows open and the error didn't show up on those. 🤷 ![image](https://github.com/microsoft/vscode/assets/22350/baad7032-018b-46f9-b526-28f8ebb7099d) cc @bpasero
1.0
Duplicate error notification from terminal process non-zero termination - <details><summary>Details</summary> <p> ``` Version: 1.80.0-insider Commit: 73eb619afb87cab4aa83a569259c891b6983c11e Date: 2023-06-19T05:56:57.953Z Electron: 22.3.11 ElectronBuildId: 21658788 Chromium: 108.0.5359.215 Node.js: 16.17.1 V8: 10.8.168.25-electron.0 OS: Darwin arm64 22.5.0 ``` </p> </details> Every now and then, exiting zsh by typing `exit` and hitting <kbd>Enter</kbd> results in an error notification. What just happened to me was that the same error notification came up in the window I did that, as well as another workspace window. I even had more workspace windows open and the error didn't show up on those. 🤷 ![image](https://github.com/microsoft/vscode/assets/22350/baad7032-018b-46f9-b526-28f8ebb7099d) cc @bpasero
process
duplicate error notification from terminal process non zero termination details version insider commit date electron electronbuildid chromium node js electron os darwin every now and then exiting zsh by typing exit and hitting enter results in an error notification what just happened to me was that the same error notification came up in the window i did that as well as another workspace window i even had more workspace windows open and the error didn t show up on those 🤷 cc bpasero
1
7,090
10,238,227,751
IssuesEvent
2019-08-19 15:26:50
hubmapconsortium/hubmap-data-portal
https://api.github.com/repos/hubmapconsortium/hubmap-data-portal
opened
Requirement specifications, Test specifications for data portal and REST api
fundamental process
We presently do not have any requirements specifications, test specs for data portal or REST api. These documents ideally help us understand/bridge the gaps between team members and understand the feature implementations better. For example: when we say Login: what are features in dataportal and REST api we intend to implement. What do we intend to test in login /logout ? Do we really **NEED** to test redirect urls from test automation tool or manual test is sufficient ? Cypress recommends not to test redirect urls, so stubs are something we should heavily rely. similarly, what are details of each feature we prefer to implement. This appears to me is applicable to all features we intend to implement.
1.0
Requirement specifications, Test specifications for data portal and REST api - We presently do not have any requirements specifications, test specs for data portal or REST api. These documents ideally help us understand/bridge the gaps between team members and understand the feature implementations better. For example: when we say Login: what are features in dataportal and REST api we intend to implement. What do we intend to test in login /logout ? Do we really **NEED** to test redirect urls from test automation tool or manual test is sufficient ? Cypress recommends not to test redirect urls, so stubs are something we should heavily rely. similarly, what are details of each feature we prefer to implement. This appears to me is applicable to all features we intend to implement.
process
requirement specifications test specifications for data portal and rest api we presently do not have any requirements specifications test specs for data portal or rest api these documents ideally help us understand bridge the gaps between team members and understand the feature implementations better for example when we say login what are features in dataportal and rest api we intend to implement what do we intend to test in login logout do we really need to test redirect urls from test automation tool or manual test is sufficient cypress recommends not to test redirect urls so stubs are something we should heavily rely similarly what are details of each feature we prefer to implement this appears to me is applicable to all features we intend to implement
1
7,550
10,675,293,145
IssuesEvent
2019-10-21 11:20:15
magento/async-import
https://api.github.com/repos/magento/async-import
closed
Make UUID for import starting not required
Contribution Day Import Process Progress: dev in progress good first issue
### Story #150 [Story] Import Process ### Description (*) For import starting, currently, UUID is a required field. See `\Magento\AsynchronousImportApi\Api\StartImportInterface` `\Magento\AsynchronousImportApi\Api\Data\ImportInterface` Should be an optional field, if not passed should be generated on Magento side If UUID is not transferred, then it have to be generated by Magento end returned back as a string from `\Magento\AsynchronousImportApi\Api\StartImportInterface`
1.0
Make UUID for import starting not required - ### Story #150 [Story] Import Process ### Description (*) For import starting, currently, UUID is a required field. See `\Magento\AsynchronousImportApi\Api\StartImportInterface` `\Magento\AsynchronousImportApi\Api\Data\ImportInterface` Should be an optional field, if not passed should be generated on Magento side If UUID is not transferred, then it have to be generated by Magento end returned back as a string from `\Magento\AsynchronousImportApi\Api\StartImportInterface`
process
make uuid for import starting not required story import process description for import starting currently uuid is a required field see magento asynchronousimportapi api startimportinterface magento asynchronousimportapi api data importinterface should be an optional field if not passed should be generated on magento side if uuid is not transferred then it have to be generated by magento end returned back as a string from magento asynchronousimportapi api startimportinterface
1
19,955
26,432,428,363
IssuesEvent
2023-01-15 00:38:14
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
[DADOS] [REMOTO] Engenheiro de Dados na [JUSBRASIL]
DESENVOLVIMENTO DE SOFTWARE REMOTO PROCESSOS INGLÊS HELP WANTED CIÊNCIA DE DADOS ENGENHARIA DE DADOS Stale
<!-- ================================================== POR FAVOR, SÓ POSTE SE A VAGA FOR PARA SALVADOR E CIDADES VIZINHAS! Use: "Desenvolvedor Front-end" ao invés de "Front-End Developer" \o/ Exemplo: `[JAVASCRIPT] [MYSQL] [NODE.JS] Desenvolvedor Front-End na [NOME DA EMPRESA]` ================================================== --> ## Descrição da vaga - Faça parte da maior plataforma jurídica do mundo e ajude-nos a diminuir o Justice GAP existente no Brasil, que é a distância entre ter um direito e fazê-lo prevalecer. - O Jusbrasil existe para conectar pessoas à justiça através de advogados e informação jurídica acessíveis a todos. Hoje temos o privilégio de observar de perto a interação de milhões de jusbrasileiros e sabemos como as pessoas querem e precisam de uma forma fácil e eficiente de "fazer valer a justiça". - Ao contrário do que muitos pensam, somos uma empresa de tecnologia, reunindo engenheiros e pessoas com background jurídico. - Estamos à procura de pessoas talentosas, que sejam apaixonadas com engenharia de dados, desenvolvimento de software e tecnologia, criativas, que primam pela qualidade dos códigos e, acima de tudo, queiram aprender e se aperfeiçoar. - Você, como Data Engineer, fará parte da equipe responsável pela estruturação e organização de dados, eventos e manutenção e evolução do nosso data lake. Nossa missão é reunir e estruturar todos os dados públicos brasileiros, e possibilitar maior transparência e melhorar a eficiência do todo o ecossistema em que a justiça está envolvida. Nós empoderamos mais de 2 milhões de usuários diários por meio de sistemas distribuídos e inteligência artificial, processando terabytes de dados - e esse é apenas o começo. ##No dia-a-dia você vai - Estruturar e organizar dados no Datalake; - Criar e gerenciar pipelines de enriquecimento e estruturação de dados; - Garantir SLA de datasets; - Dar suporte às equipes de Data Science; - Ser responsável pela governança dos dados; ## Local - Remoto ## Benefícios - Informações diretamente com o responsável/ recrutador da vaga ## Requisitos **Obrigatórios:** - Ter pelo menos 4 anos de experiência em desenvolvimento de software e/ou engenharia de dados; - Ter experiência com processamento e transformação de dados; - É capaz de implementar sistemas de grande escala; - Não ter problemas em ler conteúdo em inglês e tem compreensão auditiva (fluente é um diferencial); - Sempre está na borda do conhecimento, é uma pessoa curiosa e não ter problemas em experimentar tecnologias novas; - Fica desconfortável com códigos sem testes; **Diferenciais:** - Tem experiência com BigQuery, AWS Redshift, Hadoop, HDFS, Spark, MapReduce, Hive; - Domina algumas das tecnologias que são utilizadas no Jusbrasil como: Python, Scala, GO, ES6, Mongo, RabbitMQ, Kafka, GraphQL, Bigtable, Kubernetes, Docker, Redis, React, ElasticSearch, Google Cloud Platform, etc. - É envolvido com a comunidade de tecnologia e open source; ## Contratação - a combinar ## Nossa empresa - Para muitos a palavra justiça se resume aos tribunais, mas para nós ela deveria fazer parte das escolhas diárias de qualquer pessoa. - No meio do caminho para uma sociedade mais justa, milhões de brasileiros se deparam com um abismo entre ter um direito e torná-lo parte de suas vidas. - Somos uma startup que une Direito e Tecnologia para fazer com que a justiça ultrapasse as fronteiras dos tribunais e chegue às casas de qualquer cidadão ou cidadã, empoderando suas decisões por meio da informação. - No Brasil mais de 1,6 milhão de pessoas acessam nossa plataforma diariamente e mais de 900 mil advogados utilizam nossos produtos. ## Como se candidatar - [Clique aqui para se candidatar](https://jobs.lever.co/jusbrasil/59f756bb-5f2f-4a3b-99dc-d8eced7d4f6f)
1.0
[DADOS] [REMOTO] Engenheiro de Dados na [JUSBRASIL] - <!-- ================================================== POR FAVOR, SÓ POSTE SE A VAGA FOR PARA SALVADOR E CIDADES VIZINHAS! Use: "Desenvolvedor Front-end" ao invés de "Front-End Developer" \o/ Exemplo: `[JAVASCRIPT] [MYSQL] [NODE.JS] Desenvolvedor Front-End na [NOME DA EMPRESA]` ================================================== --> ## Descrição da vaga - Faça parte da maior plataforma jurídica do mundo e ajude-nos a diminuir o Justice GAP existente no Brasil, que é a distância entre ter um direito e fazê-lo prevalecer. - O Jusbrasil existe para conectar pessoas à justiça através de advogados e informação jurídica acessíveis a todos. Hoje temos o privilégio de observar de perto a interação de milhões de jusbrasileiros e sabemos como as pessoas querem e precisam de uma forma fácil e eficiente de "fazer valer a justiça". - Ao contrário do que muitos pensam, somos uma empresa de tecnologia, reunindo engenheiros e pessoas com background jurídico. - Estamos à procura de pessoas talentosas, que sejam apaixonadas com engenharia de dados, desenvolvimento de software e tecnologia, criativas, que primam pela qualidade dos códigos e, acima de tudo, queiram aprender e se aperfeiçoar. - Você, como Data Engineer, fará parte da equipe responsável pela estruturação e organização de dados, eventos e manutenção e evolução do nosso data lake. Nossa missão é reunir e estruturar todos os dados públicos brasileiros, e possibilitar maior transparência e melhorar a eficiência do todo o ecossistema em que a justiça está envolvida. Nós empoderamos mais de 2 milhões de usuários diários por meio de sistemas distribuídos e inteligência artificial, processando terabytes de dados - e esse é apenas o começo. ##No dia-a-dia você vai - Estruturar e organizar dados no Datalake; - Criar e gerenciar pipelines de enriquecimento e estruturação de dados; - Garantir SLA de datasets; - Dar suporte às equipes de Data Science; - Ser responsável pela governança dos dados; ## Local - Remoto ## Benefícios - Informações diretamente com o responsável/ recrutador da vaga ## Requisitos **Obrigatórios:** - Ter pelo menos 4 anos de experiência em desenvolvimento de software e/ou engenharia de dados; - Ter experiência com processamento e transformação de dados; - É capaz de implementar sistemas de grande escala; - Não ter problemas em ler conteúdo em inglês e tem compreensão auditiva (fluente é um diferencial); - Sempre está na borda do conhecimento, é uma pessoa curiosa e não ter problemas em experimentar tecnologias novas; - Fica desconfortável com códigos sem testes; **Diferenciais:** - Tem experiência com BigQuery, AWS Redshift, Hadoop, HDFS, Spark, MapReduce, Hive; - Domina algumas das tecnologias que são utilizadas no Jusbrasil como: Python, Scala, GO, ES6, Mongo, RabbitMQ, Kafka, GraphQL, Bigtable, Kubernetes, Docker, Redis, React, ElasticSearch, Google Cloud Platform, etc. - É envolvido com a comunidade de tecnologia e open source; ## Contratação - a combinar ## Nossa empresa - Para muitos a palavra justiça se resume aos tribunais, mas para nós ela deveria fazer parte das escolhas diárias de qualquer pessoa. - No meio do caminho para uma sociedade mais justa, milhões de brasileiros se deparam com um abismo entre ter um direito e torná-lo parte de suas vidas. - Somos uma startup que une Direito e Tecnologia para fazer com que a justiça ultrapasse as fronteiras dos tribunais e chegue às casas de qualquer cidadão ou cidadã, empoderando suas decisões por meio da informação. - No Brasil mais de 1,6 milhão de pessoas acessam nossa plataforma diariamente e mais de 900 mil advogados utilizam nossos produtos. ## Como se candidatar - [Clique aqui para se candidatar](https://jobs.lever.co/jusbrasil/59f756bb-5f2f-4a3b-99dc-d8eced7d4f6f)
process
engenheiro de dados na por favor só poste se a vaga for para salvador e cidades vizinhas use desenvolvedor front end ao invés de front end developer o exemplo desenvolvedor front end na descrição da vaga faça parte da maior plataforma jurídica do mundo e ajude nos a diminuir o justice gap existente no brasil que é a distância entre ter um direito e fazê lo prevalecer o jusbrasil existe para conectar pessoas à justiça através de advogados e informação jurídica acessíveis a todos hoje temos o privilégio de observar de perto a interação de milhões de jusbrasileiros e sabemos como as pessoas querem e precisam de uma forma fácil e eficiente de fazer valer a justiça ao contrário do que muitos pensam somos uma empresa de tecnologia reunindo engenheiros e pessoas com background jurídico estamos à procura de pessoas talentosas que sejam apaixonadas com engenharia de dados desenvolvimento de software e tecnologia criativas que primam pela qualidade dos códigos e acima de tudo queiram aprender e se aperfeiçoar você como data engineer fará parte da equipe responsável pela estruturação e organização de dados eventos e manutenção e evolução do nosso data lake nossa missão é reunir e estruturar todos os dados públicos brasileiros e possibilitar maior transparência e melhorar a eficiência do todo o ecossistema em que a justiça está envolvida nós empoderamos mais de milhões de usuários diários por meio de sistemas distribuídos e inteligência artificial processando terabytes de dados e esse é apenas o começo no dia a dia você vai estruturar e organizar dados no datalake criar e gerenciar pipelines de enriquecimento e estruturação de dados garantir sla de datasets dar suporte às equipes de data science ser responsável pela governança dos dados local remoto benefícios informações diretamente com o responsável recrutador da vaga requisitos obrigatórios ter pelo menos anos de experiência em desenvolvimento de software e ou engenharia de dados ter experiência com processamento e transformação de dados é capaz de implementar sistemas de grande escala não ter problemas em ler conteúdo em inglês e tem compreensão auditiva fluente é um diferencial sempre está na borda do conhecimento é uma pessoa curiosa e não ter problemas em experimentar tecnologias novas fica desconfortável com códigos sem testes diferenciais tem experiência com bigquery aws redshift hadoop hdfs spark mapreduce hive domina algumas das tecnologias que são utilizadas no jusbrasil como python scala go mongo rabbitmq kafka graphql bigtable kubernetes docker redis react elasticsearch google cloud platform etc é envolvido com a comunidade de tecnologia e open source contratação a combinar nossa empresa para muitos a palavra justiça se resume aos tribunais mas para nós ela deveria fazer parte das escolhas diárias de qualquer pessoa no meio do caminho para uma sociedade mais justa milhões de brasileiros se deparam com um abismo entre ter um direito e torná lo parte de suas vidas somos uma startup que une direito e tecnologia para fazer com que a justiça ultrapasse as fronteiras dos tribunais e chegue às casas de qualquer cidadão ou cidadã empoderando suas decisões por meio da informação no brasil mais de milhão de pessoas acessam nossa plataforma diariamente e mais de mil advogados utilizam nossos produtos como se candidatar
1
4,794
7,688,839,038
IssuesEvent
2018-05-17 10:42:43
aiidateam/aiida_core
https://api.github.com/repos/aiidateam/aiida_core
closed
calculation.copy() fails
topic/JobCalculationAndProcess type/bug
I get this traceback: ``` c.copy() --------------------------------------------------------------------------- AttributeError Traceback (most recent call last) <ipython-input-29-90d0dc129940> in <module>() ----> 1 c.copy() /home/aiida/codes/aiida_core/aiida/orm/mixins.pyc in copy(self, include_updatable_attrs) 216 if include_updatable_attrs is False: 217 for key, value in self._iter_updatable_attributes(): --> 218 clone._del_attr(key) 219 220 return clone /home/aiida/codes/aiida_core/aiida/orm/mixins.pyc in _del_attr(self, key) 205 raise ModificationNotAllowed('Cannot change the immutable attributes of a stored node') 206 --> 207 super(Sealable, self)._del_attr(key, stored_check=False) 208 209 @override /home/aiida/codes/aiida_core/aiida/orm/implementation/general/node.pyc in _del_attr(self, key, stored_check) 993 except KeyError: 994 raise AttributeError( --> 995 "DbAttribute {} does not exist".format(key)) 996 else: 997 self._del_db_attr(key) AttributeError: DbAttribute sealed does not exist ``` What is the right approach to fix this problem?
1.0
calculation.copy() fails - I get this traceback: ``` c.copy() --------------------------------------------------------------------------- AttributeError Traceback (most recent call last) <ipython-input-29-90d0dc129940> in <module>() ----> 1 c.copy() /home/aiida/codes/aiida_core/aiida/orm/mixins.pyc in copy(self, include_updatable_attrs) 216 if include_updatable_attrs is False: 217 for key, value in self._iter_updatable_attributes(): --> 218 clone._del_attr(key) 219 220 return clone /home/aiida/codes/aiida_core/aiida/orm/mixins.pyc in _del_attr(self, key) 205 raise ModificationNotAllowed('Cannot change the immutable attributes of a stored node') 206 --> 207 super(Sealable, self)._del_attr(key, stored_check=False) 208 209 @override /home/aiida/codes/aiida_core/aiida/orm/implementation/general/node.pyc in _del_attr(self, key, stored_check) 993 except KeyError: 994 raise AttributeError( --> 995 "DbAttribute {} does not exist".format(key)) 996 else: 997 self._del_db_attr(key) AttributeError: DbAttribute sealed does not exist ``` What is the right approach to fix this problem?
process
calculation copy fails i get this traceback c copy attributeerror traceback most recent call last in c copy home aiida codes aiida core aiida orm mixins pyc in copy self include updatable attrs if include updatable attrs is false for key value in self iter updatable attributes clone del attr key return clone home aiida codes aiida core aiida orm mixins pyc in del attr self key raise modificationnotallowed cannot change the immutable attributes of a stored node super sealable self del attr key stored check false override home aiida codes aiida core aiida orm implementation general node pyc in del attr self key stored check except keyerror raise attributeerror dbattribute does not exist format key else self del db attr key attributeerror dbattribute sealed does not exist what is the right approach to fix this problem
1
83,063
16,086,273,173
IssuesEvent
2021-04-26 11:38:19
joomla/joomla-cms
https://api.github.com/repos/joomla/joomla-cms
closed
[4.0] Vote: Need padding-bottom in article-view having picture
No Code Attached Yet
### What needs to be fixed As title say: Picture | Text ---- | ---- ![Ohne Titel](https://user-images.githubusercontent.com/78906980/116056933-9886e700-a67e-11eb-8ac2-ffd4b63bc308.png) | ![Ohne Titel 2](https://user-images.githubusercontent.com/78906980/116057086-bce2c380-a67e-11eb-9be8-2df226ff9bab.png) ### Why this should be fixed Look better. ### How would you fix it Append higher padding-bottom-value. ### Side Effects expected Don't know.
1.0
[4.0] Vote: Need padding-bottom in article-view having picture - ### What needs to be fixed As title say: Picture | Text ---- | ---- ![Ohne Titel](https://user-images.githubusercontent.com/78906980/116056933-9886e700-a67e-11eb-8ac2-ffd4b63bc308.png) | ![Ohne Titel 2](https://user-images.githubusercontent.com/78906980/116057086-bce2c380-a67e-11eb-9be8-2df226ff9bab.png) ### Why this should be fixed Look better. ### How would you fix it Append higher padding-bottom-value. ### Side Effects expected Don't know.
non_process
vote need padding bottom in article view having picture what needs to be fixed as title say picture text why this should be fixed look better how would you fix it append higher padding bottom value side effects expected don t know
0
19,200
25,334,864,147
IssuesEvent
2022-11-18 16:00:25
googleapis/google-api-dotnet-client
https://api.github.com/repos/googleapis/google-api-dotnet-client
closed
Update generated code that we use for integration testing
type: process priority: p2
On integration tests we use several Apiary libraries to test support and credential aspects. We have copy pasted the generated files into the integration test project at different moments, so each of the libraries has been generated with a different generator version and against different versions of the support libraries, some of these versions are fairly old. We should, at the very least, update them all to more recent versions, or right after we next publish a new generator version. Proposal: Let's link the files intead of copying, they will be kept up to date and tests will run as most users actually use the libraries. On the Google.Apis.Auth.AspNetCore3.IntegrationTests projects we use Drive and Calendar and we link the files. @jskeet if this sounds good, I'll send a PR.
1.0
Update generated code that we use for integration testing - On integration tests we use several Apiary libraries to test support and credential aspects. We have copy pasted the generated files into the integration test project at different moments, so each of the libraries has been generated with a different generator version and against different versions of the support libraries, some of these versions are fairly old. We should, at the very least, update them all to more recent versions, or right after we next publish a new generator version. Proposal: Let's link the files intead of copying, they will be kept up to date and tests will run as most users actually use the libraries. On the Google.Apis.Auth.AspNetCore3.IntegrationTests projects we use Drive and Calendar and we link the files. @jskeet if this sounds good, I'll send a PR.
process
update generated code that we use for integration testing on integration tests we use several apiary libraries to test support and credential aspects we have copy pasted the generated files into the integration test project at different moments so each of the libraries has been generated with a different generator version and against different versions of the support libraries some of these versions are fairly old we should at the very least update them all to more recent versions or right after we next publish a new generator version proposal let s link the files intead of copying they will be kept up to date and tests will run as most users actually use the libraries on the google apis auth integrationtests projects we use drive and calendar and we link the files jskeet if this sounds good i ll send a pr
1
722,183
24,853,814,903
IssuesEvent
2022-10-26 23:01:13
spiffe/spire
https://api.github.com/repos/spiffe/spire
opened
Remove x509_svid_ttl column
priority/urgent
In anticipation of introducing seperate X509-SVID and JWT-SVID TTL registration entry fields, two new columns were added to the registration entry model in the datastore (see #3174). Support was added to SPIRE core to support these fields in (#3445), which would consume these columns. However, while developing and evaluating the new support we came to the conclusion that backcompat and downgrade support became much simpler if the existing ttl column was used as it has traditionally been used, for the X509-SVID ttl. As such, the x509_svid_ttl column added in #3174 is no longer relevant. In accordance with our deprecation policies, it can be removed in 1.6.0.
1.0
Remove x509_svid_ttl column - In anticipation of introducing seperate X509-SVID and JWT-SVID TTL registration entry fields, two new columns were added to the registration entry model in the datastore (see #3174). Support was added to SPIRE core to support these fields in (#3445), which would consume these columns. However, while developing and evaluating the new support we came to the conclusion that backcompat and downgrade support became much simpler if the existing ttl column was used as it has traditionally been used, for the X509-SVID ttl. As such, the x509_svid_ttl column added in #3174 is no longer relevant. In accordance with our deprecation policies, it can be removed in 1.6.0.
non_process
remove svid ttl column in anticipation of introducing seperate svid and jwt svid ttl registration entry fields two new columns were added to the registration entry model in the datastore see support was added to spire core to support these fields in which would consume these columns however while developing and evaluating the new support we came to the conclusion that backcompat and downgrade support became much simpler if the existing ttl column was used as it has traditionally been used for the svid ttl as such the svid ttl column added in is no longer relevant in accordance with our deprecation policies it can be removed in
0
284,640
24,612,210,027
IssuesEvent
2022-10-14 23:22:12
Thy-Vipe/BeastsOfBermuda-issues
https://api.github.com/repos/Thy-Vipe/BeastsOfBermuda-issues
opened
[Bug] Music Loop Bug - Extended
bug UI public_testing
_Originally written by **Mondevu | 76561198030846914**_ Game Version: 1.1.1608 *===== System Specs ===== CPU Brand: Intel(R) Core(TM) i7-7700K CPU @ 4.20GHz Vendor: GenuineIntel GPU Brand: NVIDIA GeForce GTX 1080 GPU Driver Info: Unknown Num CPU Cores: 4 ===================* Context: **Velo** Map: Islatitania After returning to the biome in which I spawned (after dying), I can confirm that the music does not play again, even though my settings are on Loop Always.
1.0
[Bug] Music Loop Bug - Extended - _Originally written by **Mondevu | 76561198030846914**_ Game Version: 1.1.1608 *===== System Specs ===== CPU Brand: Intel(R) Core(TM) i7-7700K CPU @ 4.20GHz Vendor: GenuineIntel GPU Brand: NVIDIA GeForce GTX 1080 GPU Driver Info: Unknown Num CPU Cores: 4 ===================* Context: **Velo** Map: Islatitania After returning to the biome in which I spawned (after dying), I can confirm that the music does not play again, even though my settings are on Loop Always.
non_process
music loop bug extended originally written by mondevu game version system specs cpu brand intel r core tm cpu vendor genuineintel gpu brand nvidia geforce gtx gpu driver info unknown num cpu cores context velo map islatitania after returning to the biome in which i spawned after dying i can confirm that the music does not play again even though my settings are on loop always
0
2,935
5,919,700,342
IssuesEvent
2017-05-22 18:23:04
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
reopened
Test: System.ServiceProcess.Tests.ServiceControllerTests/StartWithArguments failed with "System.ServiceProcess.TimeoutException"
area-System.ServiceProcess test-run-portable
Opened on behalf of @Jiayili1 The test `System.ServiceProcess.Tests.ServiceControllerTests/StartWithArguments` has failed. System.ServiceProcess.TimeoutException : Time out has expired and the operation has not been completed. Stack Trace: at System.ServiceProcess.ServiceController.WaitForStatus(ServiceControllerStatus desiredStatus, TimeSpan timeout) at System.ServiceProcess.Tests.ServiceControllerTests.StartWithArguments() Build : 2.0.0 - 20170522.01 (Portable Core Tests) Failing configurations: - Windows.10.Amd64-x86 - Release Detail: https://mc.dot.net/#/product/netcore/200/source/official~2Fcorefx~2Frelease~2F2.0.0~2F/type/test~2Ffunctional~2Fportable~2Fcli~2F/build/20170522.01/workItem/System.ServiceProcess.ServiceController.Tests/analysis/xunit/System.ServiceProcess.Tests.ServiceControllerTests~2FStartWithArguments
1.0
Test: System.ServiceProcess.Tests.ServiceControllerTests/StartWithArguments failed with "System.ServiceProcess.TimeoutException" - Opened on behalf of @Jiayili1 The test `System.ServiceProcess.Tests.ServiceControllerTests/StartWithArguments` has failed. System.ServiceProcess.TimeoutException : Time out has expired and the operation has not been completed. Stack Trace: at System.ServiceProcess.ServiceController.WaitForStatus(ServiceControllerStatus desiredStatus, TimeSpan timeout) at System.ServiceProcess.Tests.ServiceControllerTests.StartWithArguments() Build : 2.0.0 - 20170522.01 (Portable Core Tests) Failing configurations: - Windows.10.Amd64-x86 - Release Detail: https://mc.dot.net/#/product/netcore/200/source/official~2Fcorefx~2Frelease~2F2.0.0~2F/type/test~2Ffunctional~2Fportable~2Fcli~2F/build/20170522.01/workItem/System.ServiceProcess.ServiceController.Tests/analysis/xunit/System.ServiceProcess.Tests.ServiceControllerTests~2FStartWithArguments
process
test system serviceprocess tests servicecontrollertests startwitharguments failed with system serviceprocess timeoutexception opened on behalf of the test system serviceprocess tests servicecontrollertests startwitharguments has failed system serviceprocess timeoutexception time out has expired and the operation has not been completed stack trace at system serviceprocess servicecontroller waitforstatus servicecontrollerstatus desiredstatus timespan timeout at system serviceprocess tests servicecontrollertests startwitharguments build portable core tests failing configurations windows release detail
1
231,641
25,525,474,808
IssuesEvent
2022-11-29 01:37:12
kapseliboi/WeiPay
https://api.github.com/repos/kapseliboi/WeiPay
reopened
CVE-2021-24036 (High) detected in follyv2016.09.26.00
security vulnerability
## CVE-2021-24036 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>follyv2016.09.26.00</b></p></summary> <p> <p>An open-source C++ library developed and used at Facebook.</p> <p>Library home page: <a href=https://github.com/facebook/folly.git>https://github.com/facebook/folly.git</a></p> <p>Found in base branch: <b>stable</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/third-party/folly-2016.09.26.00/folly/io/IOBuf.cpp</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Passing an attacker controlled size when creating an IOBuf could cause integer overflow, leading to an out of bounds write on the heap with the possibility of remote code execution. This issue affects versions of folly prior to v2021.07.22.00. This issue affects HHVM versions prior to 4.80.5, all versions between 4.81.0 and 4.102.1, all versions between 4.103.0 and 4.113.0, and versions 4.114.0, 4.115.0, 4.116.0, 4.117.0, 4.118.0 and 4.118.1. <p>Publish Date: 2021-07-23 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-24036>CVE-2021-24036</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://hhvm.com/blog/2021/07/20/security-update.html">https://hhvm.com/blog/2021/07/20/security-update.html</a></p> <p>Release Date: 2021-07-23</p> <p>Fix Resolution: HHVM - 4.80.5,4.102.2,4.113.1,4.114.1,4.115.1,4.116.1,4.117.1,4.118.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-24036 (High) detected in follyv2016.09.26.00 - ## CVE-2021-24036 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>follyv2016.09.26.00</b></p></summary> <p> <p>An open-source C++ library developed and used at Facebook.</p> <p>Library home page: <a href=https://github.com/facebook/folly.git>https://github.com/facebook/folly.git</a></p> <p>Found in base branch: <b>stable</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/third-party/folly-2016.09.26.00/folly/io/IOBuf.cpp</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Passing an attacker controlled size when creating an IOBuf could cause integer overflow, leading to an out of bounds write on the heap with the possibility of remote code execution. This issue affects versions of folly prior to v2021.07.22.00. This issue affects HHVM versions prior to 4.80.5, all versions between 4.81.0 and 4.102.1, all versions between 4.103.0 and 4.113.0, and versions 4.114.0, 4.115.0, 4.116.0, 4.117.0, 4.118.0 and 4.118.1. <p>Publish Date: 2021-07-23 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-24036>CVE-2021-24036</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://hhvm.com/blog/2021/07/20/security-update.html">https://hhvm.com/blog/2021/07/20/security-update.html</a></p> <p>Release Date: 2021-07-23</p> <p>Fix Resolution: HHVM - 4.80.5,4.102.2,4.113.1,4.114.1,4.115.1,4.116.1,4.117.1,4.118.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in cve high severity vulnerability vulnerable library an open source c library developed and used at facebook library home page a href found in base branch stable vulnerable source files third party folly folly io iobuf cpp vulnerability details passing an attacker controlled size when creating an iobuf could cause integer overflow leading to an out of bounds write on the heap with the possibility of remote code execution this issue affects versions of folly prior to this issue affects hhvm versions prior to all versions between and all versions between and and versions and publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution hhvm step up your open source security game with mend
0
13,739
16,493,836,282
IssuesEvent
2021-05-25 08:07:24
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
GO:NEW invasive hyphae formation
New term request multi-species process
background Plant infection requires an infection cell, called an appressorium, which uses a pressure-driven mechanism to breach the tough cuticle of the leaf (3, 4). Once inside plant tissue, the fungus elaborates pseudohyphalike invasive hyphae that rapidly colonize living host cells, secreting effector molecules to suppress host immunity and facilitate infection (5). we have plenty of terms for 'entry into host' but only GO:0106259 cell-to-cell migration in host for cell to cell invasion. I would therefore like to request GO:NEW invasive hyphae formation synonym cell-to-cell migration by invasive hypha (exact) synonym host tissue colonization (related) synonym plant tissue colonization (related synonym invasive growth (related, because although it is often called 'invasive growth' here we are describing the physical process rather than a size increase) is_a GO:0106259 cell-to-cell migration in host Definition The directional movement of a hyphal filament from one host cell to another. This process involves the clearance of plant-derived plasmodesmal occlusion materials, cytoskeleton based constriction of invasive hypha to traverse plasmodesmata. Septins and F-actin are reorganized into an hourglass shape at the point of maximum hyphal constriction. I know there is a term GO:0036267 invasive filamentous growth but this isn't describing invasion of cells. @CuzickA does this sound OK? Needed for PMID:29567712 instead of GO:0030448 | hyphal growth and GO:0106259 | cell-to-cell migration in host
1.0
GO:NEW invasive hyphae formation - background Plant infection requires an infection cell, called an appressorium, which uses a pressure-driven mechanism to breach the tough cuticle of the leaf (3, 4). Once inside plant tissue, the fungus elaborates pseudohyphalike invasive hyphae that rapidly colonize living host cells, secreting effector molecules to suppress host immunity and facilitate infection (5). we have plenty of terms for 'entry into host' but only GO:0106259 cell-to-cell migration in host for cell to cell invasion. I would therefore like to request GO:NEW invasive hyphae formation synonym cell-to-cell migration by invasive hypha (exact) synonym host tissue colonization (related) synonym plant tissue colonization (related synonym invasive growth (related, because although it is often called 'invasive growth' here we are describing the physical process rather than a size increase) is_a GO:0106259 cell-to-cell migration in host Definition The directional movement of a hyphal filament from one host cell to another. This process involves the clearance of plant-derived plasmodesmal occlusion materials, cytoskeleton based constriction of invasive hypha to traverse plasmodesmata. Septins and F-actin are reorganized into an hourglass shape at the point of maximum hyphal constriction. I know there is a term GO:0036267 invasive filamentous growth but this isn't describing invasion of cells. @CuzickA does this sound OK? Needed for PMID:29567712 instead of GO:0030448 | hyphal growth and GO:0106259 | cell-to-cell migration in host
process
go new invasive hyphae formation background plant infection requires an infection cell called an appressorium which uses a pressure driven mechanism to breach the tough cuticle of the leaf once inside plant tissue the fungus elaborates pseudohyphalike invasive hyphae that rapidly colonize living host cells secreting effector molecules to suppress host immunity and facilitate infection we have plenty of terms for entry into host but only go cell to cell migration in host for cell to cell invasion i would therefore like to request go new invasive hyphae formation synonym cell to cell migration by invasive hypha exact synonym host tissue colonization related synonym plant tissue colonization related synonym invasive growth related because although it is often called invasive growth here we are describing the physical process rather than a size increase is a go cell to cell migration in host definition the directional movement of a hyphal filament from one host cell to another this process involves the clearance of plant derived plasmodesmal occlusion materials cytoskeleton based constriction of invasive hypha to traverse plasmodesmata septins and f actin are reorganized into an hourglass shape at the point of maximum hyphal constriction i know there is a term go invasive filamentous growth but this isn t describing invasion of cells cuzicka does this sound ok needed for pmid instead of go hyphal growth and go cell to cell migration in host
1
19,458
25,745,616,869
IssuesEvent
2022-12-08 09:38:26
dotnet/dotnet-api-docs
https://api.github.com/repos/dotnet/dotnet-api-docs
closed
Is the OnStart method document of servicebase incorrect
untriaged Pri3 area-System.ServiceProcess
look this code snippet, ```c# Environment.GetCommandLineArgs() ``` It's can not get registry,but in the [OnStart](https://docs.microsoft.com/en-us/dotnet/api/system.serviceprocess.servicebase.onstart?view=dotnet-plat-ext-5.0#System_ServiceProcess_ServiceBase_OnStart_System_String___) document that we can get registry infomation.
1.0
Is the OnStart method document of servicebase incorrect - look this code snippet, ```c# Environment.GetCommandLineArgs() ``` It's can not get registry,but in the [OnStart](https://docs.microsoft.com/en-us/dotnet/api/system.serviceprocess.servicebase.onstart?view=dotnet-plat-ext-5.0#System_ServiceProcess_ServiceBase_OnStart_System_String___) document that we can get registry infomation.
process
is the onstart method document of servicebase incorrect look this code snippet, c environment getcommandlineargs it s can not get registry but in the document that we can get registry infomation
1
451,949
13,044,049,369
IssuesEvent
2020-07-29 03:26:22
33cn/plugin
https://api.github.com/repos/33cn/plugin
closed
mempool 删除大量交易缓存速度较慢
high priority
4个节点组成 tendermint 集群,每个节点依次循环打包区块,其中 mempool 排序策略为 timeline 方式 当每个节点每秒钟收到1000笔交易时,出现 节点需要等待4,5秒才能获取到新的交易 初步判断是节点打包区块后通知 mempool 删除其交易缓存的速度较慢
1.0
mempool 删除大量交易缓存速度较慢 - 4个节点组成 tendermint 集群,每个节点依次循环打包区块,其中 mempool 排序策略为 timeline 方式 当每个节点每秒钟收到1000笔交易时,出现 节点需要等待4,5秒才能获取到新的交易 初步判断是节点打包区块后通知 mempool 删除其交易缓存的速度较慢
non_process
mempool 删除大量交易缓存速度较慢 tendermint 集群,每个节点依次循环打包区块,其中 mempool 排序策略为 timeline 方式 ,出现 , 初步判断是节点打包区块后通知 mempool 删除其交易缓存的速度较慢
0
111,287
11,728,472,515
IssuesEvent
2020-03-10 17:36:37
davorpa/gpul-labs-android-scissors-game
https://api.github.com/repos/davorpa/gpul-labs-android-scissors-game
opened
Update README with library used version
documentation good first issue
Update README.md to reflect versions of Andoid Tools used Frameworks ...
1.0
Update README with library used version - Update README.md to reflect versions of Andoid Tools used Frameworks ...
non_process
update readme with library used version update readme md to reflect versions of andoid tools used frameworks
0
253,912
27,337,574,113
IssuesEvent
2023-02-26 12:08:22
DEV-REPO-URIEL/Example
https://api.github.com/repos/DEV-REPO-URIEL/Example
opened
babel-loader-6.4.1.tgz: 1 vulnerabilities (highest severity is: 9.8)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>babel-loader-6.4.1.tgz</b></p></summary> <p></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/loader-utils/package.json</p> <p> <p>Found in HEAD commit: <a href="https://github.com/DEV-REPO-URIEL/Example/commit/98862553de1a203e077fe9dcfa28aaab87be3937">98862553de1a203e077fe9dcfa28aaab87be3937</a></p></details> #### <img src='https://whitesource-resources.whitesourcesoftware.com/suggestedVersion.png' width=19 height=20> Mend has checked all newer package trees, and you are on the least vulnerable package! #### Please note: There might be a version that explicitly solves one or more of the vulnerabilities listed below, but we do not recommend it. For more info about the optional fixes, check the section “Details” below. ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (babel-loader version) | Fix PR available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2022-37601](https://www.mend.io/vulnerability-database/CVE-2022-37601) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | loader-utils-0.2.17.tgz | Transitive | N/A* | &#10060; | <p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-37601</summary> ### Vulnerable Library - <b>loader-utils-0.2.17.tgz</b></p> <p>utils for webpack loaders</p> <p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-0.2.17.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-0.2.17.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/loader-utils/package.json</p> <p> Dependency Hierarchy: - babel-loader-6.4.1.tgz (Root Library) - :x: **loader-utils-0.2.17.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/DEV-REPO-URIEL/Example/commit/98862553de1a203e077fe9dcfa28aaab87be3937">98862553de1a203e077fe9dcfa28aaab87be3937</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> Prototype pollution vulnerability in function parseQuery in parseQuery.js in webpack loader-utils 2.0.0 via the name variable in parseQuery.js. <p>Publish Date: 2022-10-12 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-37601>CVE-2022-37601</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-10-12</p> <p>Fix Resolution: loader-utils - v2.0.0</p> </p> <p></p> </details>
True
babel-loader-6.4.1.tgz: 1 vulnerabilities (highest severity is: 9.8) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>babel-loader-6.4.1.tgz</b></p></summary> <p></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/loader-utils/package.json</p> <p> <p>Found in HEAD commit: <a href="https://github.com/DEV-REPO-URIEL/Example/commit/98862553de1a203e077fe9dcfa28aaab87be3937">98862553de1a203e077fe9dcfa28aaab87be3937</a></p></details> #### <img src='https://whitesource-resources.whitesourcesoftware.com/suggestedVersion.png' width=19 height=20> Mend has checked all newer package trees, and you are on the least vulnerable package! #### Please note: There might be a version that explicitly solves one or more of the vulnerabilities listed below, but we do not recommend it. For more info about the optional fixes, check the section “Details” below. ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (babel-loader version) | Fix PR available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2022-37601](https://www.mend.io/vulnerability-database/CVE-2022-37601) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | loader-utils-0.2.17.tgz | Transitive | N/A* | &#10060; | <p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-37601</summary> ### Vulnerable Library - <b>loader-utils-0.2.17.tgz</b></p> <p>utils for webpack loaders</p> <p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-0.2.17.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-0.2.17.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/loader-utils/package.json</p> <p> Dependency Hierarchy: - babel-loader-6.4.1.tgz (Root Library) - :x: **loader-utils-0.2.17.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/DEV-REPO-URIEL/Example/commit/98862553de1a203e077fe9dcfa28aaab87be3937">98862553de1a203e077fe9dcfa28aaab87be3937</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> Prototype pollution vulnerability in function parseQuery in parseQuery.js in webpack loader-utils 2.0.0 via the name variable in parseQuery.js. <p>Publish Date: 2022-10-12 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-37601>CVE-2022-37601</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-10-12</p> <p>Fix Resolution: loader-utils - v2.0.0</p> </p> <p></p> </details>
non_process
babel loader tgz vulnerabilities highest severity is vulnerable library babel loader tgz path to dependency file package json path to vulnerable library node modules loader utils package json found in head commit a href mend has checked all newer package trees and you are on the least vulnerable package please note there might be a version that explicitly solves one or more of the vulnerabilities listed below but we do not recommend it for more info about the optional fixes check the section “details” below vulnerabilities cve severity cvss dependency type fixed in babel loader version fix pr available high loader utils tgz transitive n a for some transitive vulnerabilities there is no version of direct dependency with a fix check the section details below to see if there is a version of transitive dependency where vulnerability is fixed details cve vulnerable library loader utils tgz utils for webpack loaders library home page a href path to dependency file package json path to vulnerable library node modules loader utils package json dependency hierarchy babel loader tgz root library x loader utils tgz vulnerable library found in head commit a href found in base branch main vulnerability details prototype pollution vulnerability in function parsequery in parsequery js in webpack loader utils via the name variable in parsequery js publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution loader utils
0
55,594
14,579,476,740
IssuesEvent
2020-12-18 07:23:00
PowerDNS/pdns
https://api.github.com/repos/PowerDNS/pdns
closed
broken link in recursor hooks documentation
defect docs easy rec
- Program: Recursor - Issue type: Bug report ### Short description ./pdns/recursordist/docs/lua-scripting/hooks.rst https://docs.powerdns.com/recursor/lua-scripting/hooks.html > Once a script is loaded, PowerDNS looks for several functions in the loaded script. All of these functions are optional. The word 'functions' is a broken link. It's entirely possible it's supposed to be self-referential but may not be super-useful. ### Environment Web documentation. ### Steps to reproduce 1. load https://docs.powerdns.com/recursor/lua-scripting/hooks.html in a browser of your choice 2. click the word 'functions' in this section: https://docs.powerdns.com/recursor/lua-scripting/hooks.html#writing-lua-powerdns-recursor-scripts 3. weep at 404 Thanks
1.0
broken link in recursor hooks documentation - - Program: Recursor - Issue type: Bug report ### Short description ./pdns/recursordist/docs/lua-scripting/hooks.rst https://docs.powerdns.com/recursor/lua-scripting/hooks.html > Once a script is loaded, PowerDNS looks for several functions in the loaded script. All of these functions are optional. The word 'functions' is a broken link. It's entirely possible it's supposed to be self-referential but may not be super-useful. ### Environment Web documentation. ### Steps to reproduce 1. load https://docs.powerdns.com/recursor/lua-scripting/hooks.html in a browser of your choice 2. click the word 'functions' in this section: https://docs.powerdns.com/recursor/lua-scripting/hooks.html#writing-lua-powerdns-recursor-scripts 3. weep at 404 Thanks
non_process
broken link in recursor hooks documentation program recursor issue type bug report short description pdns recursordist docs lua scripting hooks rst once a script is loaded powerdns looks for several functions in the loaded script all of these functions are optional the word functions is a broken link it s entirely possible it s supposed to be self referential but may not be super useful environment web documentation steps to reproduce load in a browser of your choice click the word functions in this section weep at thanks
0
14,566
17,689,938,941
IssuesEvent
2021-08-24 08:42:54
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Add support for stopping WVD Hosts
automation/svc triaged cxp awaiting-product-team-response product-feedback process-automation/subsvc Pri1
I'm not sure if this is the right place, but I think this solution can be easily expanded to be able to gracefully shutdown WVD hosts, as users logged in needs to be logged off first before the VM can be shut down, for the users to be able to login to another WVD host. I have prepared a script which does this, also, it works for me, just when adding the below code on the line 48 of the ScheduledStartStop_Child runbook (I'm not using the sequenced runbooks though): $pools = get-azwvdHostPool Write-Output "Checking if VM $VMName is part of a WVD host pool" Foreach ($pool in $pools) { $shost = Get-AzWvdSessionHost -HostPoolName $pool.Name -ResourceGroupName $ResourceGroupName | ? {$_.Name -like "*$VMName*"} } if ($shost -ne $null) { Write-Output "VM $VMName is part of a WVD host pool $pool.Name" $sessions = Get-AzWvdUserSession -HostPoolName $pool.Name -ResourceGroupName $ResourceGroupName | ? {$_.Name -like "*$VMName*"} if ($sessions -ne $null) { Write-Output "There are $sessions.Count session(s), logging them off from a VWD host $VMName" $sessions | % { Remove-AzWvdUserSession -SessionHostName $shost.Name.Split("/")[-1] -HostPoolName $pool.Name -ResourceGroupName WVD -Id $_.Id.Split("/")[-1] } sleep 60 } else { Write-Output "VWD host $VMName has no logged in users" } } else { Write-Output "VM $VMName is not part of a WVD HostPool" } ``` --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 225c9d05-83dd-b006-0025-3753f5ab25bf * Version Independent ID: 9eecef0c-b1cb-1136-faf7-542214492096 * Content: [Azure Automation Start/Stop VMs during off-hours overview](https://docs.microsoft.com/en-us/azure/automation/automation-solution-vm-management) * Content Source: [articles/automation/automation-solution-vm-management.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/automation-solution-vm-management.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @MGoedtel * Microsoft Alias: **magoedte**
1.0
Add support for stopping WVD Hosts - I'm not sure if this is the right place, but I think this solution can be easily expanded to be able to gracefully shutdown WVD hosts, as users logged in needs to be logged off first before the VM can be shut down, for the users to be able to login to another WVD host. I have prepared a script which does this, also, it works for me, just when adding the below code on the line 48 of the ScheduledStartStop_Child runbook (I'm not using the sequenced runbooks though): $pools = get-azwvdHostPool Write-Output "Checking if VM $VMName is part of a WVD host pool" Foreach ($pool in $pools) { $shost = Get-AzWvdSessionHost -HostPoolName $pool.Name -ResourceGroupName $ResourceGroupName | ? {$_.Name -like "*$VMName*"} } if ($shost -ne $null) { Write-Output "VM $VMName is part of a WVD host pool $pool.Name" $sessions = Get-AzWvdUserSession -HostPoolName $pool.Name -ResourceGroupName $ResourceGroupName | ? {$_.Name -like "*$VMName*"} if ($sessions -ne $null) { Write-Output "There are $sessions.Count session(s), logging them off from a VWD host $VMName" $sessions | % { Remove-AzWvdUserSession -SessionHostName $shost.Name.Split("/")[-1] -HostPoolName $pool.Name -ResourceGroupName WVD -Id $_.Id.Split("/")[-1] } sleep 60 } else { Write-Output "VWD host $VMName has no logged in users" } } else { Write-Output "VM $VMName is not part of a WVD HostPool" } ``` --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 225c9d05-83dd-b006-0025-3753f5ab25bf * Version Independent ID: 9eecef0c-b1cb-1136-faf7-542214492096 * Content: [Azure Automation Start/Stop VMs during off-hours overview](https://docs.microsoft.com/en-us/azure/automation/automation-solution-vm-management) * Content Source: [articles/automation/automation-solution-vm-management.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/automation/automation-solution-vm-management.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @MGoedtel * Microsoft Alias: **magoedte**
process
add support for stopping wvd hosts i m not sure if this is the right place but i think this solution can be easily expanded to be able to gracefully shutdown wvd hosts as users logged in needs to be logged off first before the vm can be shut down for the users to be able to login to another wvd host i have prepared a script which does this also it works for me just when adding the below code on the line of the scheduledstartstop child runbook i m not using the sequenced runbooks though pools get azwvdhostpool write output checking if vm vmname is part of a wvd host pool foreach pool in pools shost get azwvdsessionhost hostpoolname pool name resourcegroupname resourcegroupname name like vmname if shost ne null write output vm vmname is part of a wvd host pool pool name sessions get azwvdusersession hostpoolname pool name resourcegroupname resourcegroupname name like vmname if sessions ne null write output there are sessions count session s logging them off from a vwd host vmname sessions remove azwvdusersession sessionhostname shost name split hostpoolname pool name resourcegroupname wvd id id split sleep else write output vwd host vmname has no logged in users else write output vm vmname is not part of a wvd hostpool document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login mgoedtel microsoft alias magoedte
1
560,675
16,601,490,561
IssuesEvent
2021-06-01 20:08:19
google/ground-platform
https://api.github.com/repos/google/ground-platform
opened
[Feature] Layer labels for polygons and points
priority: p2 type: feature request web
Currently, only a color designates and distinguishes one layer from another. While this works, it is slightly inconvenient to have to recollect color associations when viewing the map. Supporting layer labels on points and polygons would make it easier to quickly identify the layer to which a given feature belongs.
1.0
[Feature] Layer labels for polygons and points - Currently, only a color designates and distinguishes one layer from another. While this works, it is slightly inconvenient to have to recollect color associations when viewing the map. Supporting layer labels on points and polygons would make it easier to quickly identify the layer to which a given feature belongs.
non_process
layer labels for polygons and points currently only a color designates and distinguishes one layer from another while this works it is slightly inconvenient to have to recollect color associations when viewing the map supporting layer labels on points and polygons would make it easier to quickly identify the layer to which a given feature belongs
0
293,680
25,316,424,280
IssuesEvent
2022-11-17 22:05:32
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
closed
Extend detection for ad ref code
QA/Yes release-notes/exclude QA/Test-Plan-Specified OS/Android
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!-- Provide a brief description of the issue --> This a follow up for https://github.com/brave/brave-browser/issues/25265 We need to extend detection for ad ref code.
1.0
Extend detection for ad ref code - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!-- Provide a brief description of the issue --> This a follow up for https://github.com/brave/brave-browser/issues/25265 We need to extend detection for ad ref code.
non_process
extend detection for ad ref code have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description this a follow up for we need to extend detection for ad ref code
0
17,241
22,970,796,981
IssuesEvent
2022-07-20 02:46:10
zotero/zotero
https://api.github.com/repos/zotero/zotero
opened
Add a dropmarker to Quick Format citation bubbles
Word Processor Integration
https://twitter.com/grace_lao/status/1549544846913933313 I've been thinking for a while that we need to make the popup more obvious. People in the forums pretty regularly have no idea that there's more to the dialog. We probably want a light-blueish triangle on the right-hand side. Will slightly ruin the asymmetry of the bubbles, but I think it's worth it for the increased usability.
1.0
Add a dropmarker to Quick Format citation bubbles - https://twitter.com/grace_lao/status/1549544846913933313 I've been thinking for a while that we need to make the popup more obvious. People in the forums pretty regularly have no idea that there's more to the dialog. We probably want a light-blueish triangle on the right-hand side. Will slightly ruin the asymmetry of the bubbles, but I think it's worth it for the increased usability.
process
add a dropmarker to quick format citation bubbles i ve been thinking for a while that we need to make the popup more obvious people in the forums pretty regularly have no idea that there s more to the dialog we probably want a light blueish triangle on the right hand side will slightly ruin the asymmetry of the bubbles but i think it s worth it for the increased usability
1
54,355
11,219,650,674
IssuesEvent
2020-01-07 14:19:54
hazelcast/hazelcast-jet
https://api.github.com/repos/hazelcast/hazelcast-jet
closed
hello-world example does not print anything if submitted through CLI
code-samples defect
In case when `hello-world` example jar from https://github.com/hazelcast/hazelcast-jet/tree/master/examples/hello-world is submitted through jet CLI then expected result is not printed. It seems it is caused by missing log4j properties. We should provide and documented simple way how to run execute this code sample properly in https://github.com/hazelcast/hazelcast-jet/blob/d03eb34092aa7c66c47647967c24d42e72c66901/examples/hello-world/src/main/java/com/hazelcast/jet/examples/helloworld/HelloWorld.java#L38. Current output is following: ``` log4j:WARN No appenders could be found for logger (com.hazelcast.jet.examples.helloworld.HelloWorld). log4j:WARN Please initialize the log4j system properly. log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for more info. ``` It seems to be related to removing log4j from `jet` CLI in https://github.com/hazelcast/hazelcast-jet/commit/9051ec1ffb55d6504325d6945364397591fe3c0e
1.0
hello-world example does not print anything if submitted through CLI - In case when `hello-world` example jar from https://github.com/hazelcast/hazelcast-jet/tree/master/examples/hello-world is submitted through jet CLI then expected result is not printed. It seems it is caused by missing log4j properties. We should provide and documented simple way how to run execute this code sample properly in https://github.com/hazelcast/hazelcast-jet/blob/d03eb34092aa7c66c47647967c24d42e72c66901/examples/hello-world/src/main/java/com/hazelcast/jet/examples/helloworld/HelloWorld.java#L38. Current output is following: ``` log4j:WARN No appenders could be found for logger (com.hazelcast.jet.examples.helloworld.HelloWorld). log4j:WARN Please initialize the log4j system properly. log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for more info. ``` It seems to be related to removing log4j from `jet` CLI in https://github.com/hazelcast/hazelcast-jet/commit/9051ec1ffb55d6504325d6945364397591fe3c0e
non_process
hello world example does not print anything if submitted through cli in case when hello world example jar from is submitted through jet cli then expected result is not printed it seems it is caused by missing properties we should provide and documented simple way how to run execute this code sample properly in current output is following warn no appenders could be found for logger com hazelcast jet examples helloworld helloworld warn please initialize the system properly warn see for more info it seems to be related to removing from jet cli in
0
8,365
11,519,059,461
IssuesEvent
2020-02-14 11:54:52
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
Obsoletion of unused symbiont 'growth' terms
multi-species process obsoletion
Hello, These terms combine a process and a location, so they will be obsoleted: * GO:0044151 growth of host on or near symbiont surface * GO:0044133 growth of symbiont on or near host (2 EXP ASAP @pgaudet) * GO:0044135 growth of symbiont on or near host phyllosphere * GO:0044137 growth of symbiont on or near host rhizosphere
1.0
Obsoletion of unused symbiont 'growth' terms - Hello, These terms combine a process and a location, so they will be obsoleted: * GO:0044151 growth of host on or near symbiont surface * GO:0044133 growth of symbiont on or near host (2 EXP ASAP @pgaudet) * GO:0044135 growth of symbiont on or near host phyllosphere * GO:0044137 growth of symbiont on or near host rhizosphere
process
obsoletion of unused symbiont growth terms hello these terms combine a process and a location so they will be obsoleted go growth of host on or near symbiont surface go growth of symbiont on or near host exp asap pgaudet go growth of symbiont on or near host phyllosphere go growth of symbiont on or near host rhizosphere
1
5,532
8,391,255,487
IssuesEvent
2018-10-09 14:36:28
kiwicom/orbit-components
https://api.github.com/repos/kiwicom/orbit-components
closed
Select: help/error is positioned to whichever parent element has position relative
bug processing
## Expected Behavior Help/error to be underneath the input field. ## Current Behavior help/error is now placed to the first parent element that has position relative on itself. ![image](https://user-images.githubusercontent.com/16268406/46586370-e36b8f80-ca7d-11e8-80fc-a2ed676ee9ad.png) ## Possible Solution 1) Remove position absolute or 2) put position relative on label, but then also include margin-bottom on the label so that the element underneath is not overlapping with the help/error ## Context (Environment) using Orbit 0.13.0
1.0
Select: help/error is positioned to whichever parent element has position relative - ## Expected Behavior Help/error to be underneath the input field. ## Current Behavior help/error is now placed to the first parent element that has position relative on itself. ![image](https://user-images.githubusercontent.com/16268406/46586370-e36b8f80-ca7d-11e8-80fc-a2ed676ee9ad.png) ## Possible Solution 1) Remove position absolute or 2) put position relative on label, but then also include margin-bottom on the label so that the element underneath is not overlapping with the help/error ## Context (Environment) using Orbit 0.13.0
process
select help error is positioned to whichever parent element has position relative expected behavior help error to be underneath the input field current behavior help error is now placed to the first parent element that has position relative on itself possible solution remove position absolute or put position relative on label but then also include margin bottom on the label so that the element underneath is not overlapping with the help error context environment using orbit
1
2,528
5,288,721,021
IssuesEvent
2017-02-08 15:47:01
symfony/symfony
https://api.github.com/repos/symfony/symfony
closed
[Process] Windows escaping is not consistent
Enhancement Process
After digging more on Windows escaping, I realised some things: - We forbid environment variable expansion by escaping `%APPDATA%` to `^%"APPDATA"^%` - We explicitly ask for variable expansion at runtime (running the command line with the [`/V:ON`](https://github.com/symfony/symfony/blob/2.3/src/Symfony/Component/Process/Process.php#L235) flag). Running a command containing `!APPDATA!` will be escaped and expanded (our previous rule is easily overriden) - On platform that are not windows, we use strong escaping that prevents any variable expansion (`$PATH` will be escaped to `'$PATH'` that is not interpreted as the current PATH) We have three possibilities: - Keep this behavior as this. - Prefer a consistent API and use a strong escaping strategy everywhere, but it would result in a BC break (see #8975). - Allow environment variable expansion and escape `%APPDATA%` to `"%APPDATA%"` Any thoughts about this ?
1.0
[Process] Windows escaping is not consistent - After digging more on Windows escaping, I realised some things: - We forbid environment variable expansion by escaping `%APPDATA%` to `^%"APPDATA"^%` - We explicitly ask for variable expansion at runtime (running the command line with the [`/V:ON`](https://github.com/symfony/symfony/blob/2.3/src/Symfony/Component/Process/Process.php#L235) flag). Running a command containing `!APPDATA!` will be escaped and expanded (our previous rule is easily overriden) - On platform that are not windows, we use strong escaping that prevents any variable expansion (`$PATH` will be escaped to `'$PATH'` that is not interpreted as the current PATH) We have three possibilities: - Keep this behavior as this. - Prefer a consistent API and use a strong escaping strategy everywhere, but it would result in a BC break (see #8975). - Allow environment variable expansion and escape `%APPDATA%` to `"%APPDATA%"` Any thoughts about this ?
process
windows escaping is not consistent after digging more on windows escaping i realised some things we forbid environment variable expansion by escaping appdata to appdata we explicitly ask for variable expansion at runtime running the command line with the flag running a command containing appdata will be escaped and expanded our previous rule is easily overriden on platform that are not windows we use strong escaping that prevents any variable expansion path will be escaped to path that is not interpreted as the current path we have three possibilities keep this behavior as this prefer a consistent api and use a strong escaping strategy everywhere but it would result in a bc break see allow environment variable expansion and escape appdata to appdata any thoughts about this
1
15,358
19,530,534,349
IssuesEvent
2021-12-30 15:56:48
MikeKSmith/The_Lazy_Producer
https://api.github.com/repos/MikeKSmith/The_Lazy_Producer
opened
Discuss automation
process
How do we prevent ambient and generative music and the sounds used from becoming boring? Using modulation plugins and sources.
1.0
Discuss automation - How do we prevent ambient and generative music and the sounds used from becoming boring? Using modulation plugins and sources.
process
discuss automation how do we prevent ambient and generative music and the sounds used from becoming boring using modulation plugins and sources
1
108,378
9,303,754,574
IssuesEvent
2019-03-24 19:51:14
strongbox/strongbox
https://api.github.com/repos/strongbox/strongbox
opened
Upgrade the version of Gradle in the strongbox-web-integration-tests to 5.3, or higher
good first issue help wanted testing
# Task Description We should investigate what needs to be done in order to support newer versions of Gradle (>= 5.3) in the [`strongbox-web-integration-tests`][strongbox-web-integration-tests]. # Acceptance Test * [ ] The tests in the [`strongbox-web-integration-tests`][strongbox-web-integration-tests] still run properly. # Help * [Our chat](https://chat.carlspring.org/) * Points of contact: * @carlspring * @sbespalov * @fuss86 [strongbox-web-integration-tests]: https://github.com/strongbox/strongbox-web-integration-tests/
1.0
Upgrade the version of Gradle in the strongbox-web-integration-tests to 5.3, or higher - # Task Description We should investigate what needs to be done in order to support newer versions of Gradle (>= 5.3) in the [`strongbox-web-integration-tests`][strongbox-web-integration-tests]. # Acceptance Test * [ ] The tests in the [`strongbox-web-integration-tests`][strongbox-web-integration-tests] still run properly. # Help * [Our chat](https://chat.carlspring.org/) * Points of contact: * @carlspring * @sbespalov * @fuss86 [strongbox-web-integration-tests]: https://github.com/strongbox/strongbox-web-integration-tests/
non_process
upgrade the version of gradle in the strongbox web integration tests to or higher task description we should investigate what needs to be done in order to support newer versions of gradle in the acceptance test the tests in the still run properly help points of contact carlspring sbespalov
0
10,414
13,208,207,117
IssuesEvent
2020-08-15 03:06:19
kubeflow/website
https://api.github.com/repos/kubeflow/website
closed
About pages need OWNERs file
area/community area/docs kind/process priority/p0
https://github.com/kubeflow/website/tree/master/content/en/docs/about This directory needs an OWNERs file. It also looks like the pages are due for a refresh to get rid of the outdated banner. /cc @Bobgy @joeliedtke @jbottum @8bitmp3
1.0
About pages need OWNERs file - https://github.com/kubeflow/website/tree/master/content/en/docs/about This directory needs an OWNERs file. It also looks like the pages are due for a refresh to get rid of the outdated banner. /cc @Bobgy @joeliedtke @jbottum @8bitmp3
process
about pages need owners file this directory needs an owners file it also looks like the pages are due for a refresh to get rid of the outdated banner cc bobgy joeliedtke jbottum
1
4,536
7,373,416,101
IssuesEvent
2018-03-13 17:12:56
wpninjas/ninja-forms
https://api.github.com/repos/wpninjas/ninja-forms
closed
The Star Rating field is not being caught by the required field validation
DIFFICULTY: Easy FRONT: Processing PRIORITY: Low VALUE: Painless
The Star Rating field is not being caught by the required field validation in Ninja Forms 3.1.4
1.0
The Star Rating field is not being caught by the required field validation - The Star Rating field is not being caught by the required field validation in Ninja Forms 3.1.4
process
the star rating field is not being caught by the required field validation the star rating field is not being caught by the required field validation in ninja forms
1
7,597
10,707,488,323
IssuesEvent
2019-10-24 17:34:58
SpongePowered/Mixin
https://api.github.com/repos/SpongePowered/Mixin
closed
Allow warning to be suppressed for public Mixin#targets when target is in the default package
annotation processor core enhancement
When I used default package class in Mixin's targets, I get this warning. We can not import default package so please allow using default package in Mixin `warning: Mixin target ~~~ is public and must be specified in value @Mixin(targets = {"~~~"})`
1.0
Allow warning to be suppressed for public Mixin#targets when target is in the default package - When I used default package class in Mixin's targets, I get this warning. We can not import default package so please allow using default package in Mixin `warning: Mixin target ~~~ is public and must be specified in value @Mixin(targets = {"~~~"})`
process
allow warning to be suppressed for public mixin targets when target is in the default package when i used default package class in mixin s targets i get this warning we can not import default package so please allow using default package in mixin warning mixin target is public and must be specified in value mixin targets
1
694,593
23,820,196,247
IssuesEvent
2022-09-05 10:29:16
ntop/ntopng
https://api.github.com/repos/ntop/ntopng
closed
Broken Layout in Manage Users
Low-Priority Bug
![image](https://user-images.githubusercontent.com/4493366/188424573-a414ed0b-99ba-4694-ba9c-5aa5f549ec6f.png) The language drop down menu is empty and too large. Some spaces are missing.
1.0
Broken Layout in Manage Users - ![image](https://user-images.githubusercontent.com/4493366/188424573-a414ed0b-99ba-4694-ba9c-5aa5f549ec6f.png) The language drop down menu is empty and too large. Some spaces are missing.
non_process
broken layout in manage users the language drop down menu is empty and too large some spaces are missing
0
546,050
15,983,345,706
IssuesEvent
2021-04-18 08:46:15
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
closed
[Package version resolution] Give priority to a user-specified version when picking a dependency package version
Area/ProjectAPI Priority/Blocker Team/Tooling Type/Bug
Consider the following example: ``` A --> B (1.2.0-alpha) --> C (1.0.0-alpha) A --> C (version is not specified) ``` Assume that there exists a new version of `C` in Ballerina central. At the moment, the package version resolution algorithm picks the latest compatible version, even though `1.0.0-alpha` version is available as a transitive dependency. As per the [design](https://github.com/ballerina-platform/ballerina-spec/issues/667#issuecomment-816310162), the algorithm should pick `1.0.0-alpha`
1.0
[Package version resolution] Give priority to a user-specified version when picking a dependency package version - Consider the following example: ``` A --> B (1.2.0-alpha) --> C (1.0.0-alpha) A --> C (version is not specified) ``` Assume that there exists a new version of `C` in Ballerina central. At the moment, the package version resolution algorithm picks the latest compatible version, even though `1.0.0-alpha` version is available as a transitive dependency. As per the [design](https://github.com/ballerina-platform/ballerina-spec/issues/667#issuecomment-816310162), the algorithm should pick `1.0.0-alpha`
non_process
give priority to a user specified version when picking a dependency package version consider the following example a b alpha c alpha a c version is not specified assume that there exists a new version of c in ballerina central at the moment the package version resolution algorithm picks the latest compatible version even though alpha version is available as a transitive dependency as per the the algorithm should pick alpha
0
20,085
26,585,231,976
IssuesEvent
2023-01-22 23:13:15
pb866/Kimai.jl
https://api.github.com/repos/pb866/Kimai.jl
opened
Reset counter for sick leave at the end of the year
enhancement data processing
- [ ] Reset counter for sick days at the beginning of new year - [ ] Save history of sick days
1.0
Reset counter for sick leave at the end of the year - - [ ] Reset counter for sick days at the beginning of new year - [ ] Save history of sick days
process
reset counter for sick leave at the end of the year reset counter for sick days at the beginning of new year save history of sick days
1
140,268
5,399,514,622
IssuesEvent
2017-02-27 19:38:41
coreos/bugs
https://api.github.com/repos/coreos/bugs
closed
Ignition not working on OpenStack
area/usability component/ignition kind/bug priority/P0 team/os
## Bug ## ### Container Linux Version ### ``` core@localhost ~ $ cat /etc/os-release NAME="Container Linux by CoreOS" ID=coreos VERSION=1235.9.0 VERSION_ID=1235.9.0 BUILD_ID=2017-02-02-0235 PRETTY_NAME="Container Linux by CoreOS 1235.9.0 (Ladybug)" ANSI_COLOR="38;5;75" HOME_URL="https://coreos.com/" BUG_REPORT_URL="https://github.com/coreos/bugs/issues" ``` ### Environment ### Openstack - https://vexxhost.com/ ### Expected Behavior ### Ignition configs should work. ### Actual Behavior ### Ignition configs don't work. ### Reproduction Steps ### https://gist.github.com/s-urbaniak/8108dfa3269da858c489f8cdf38d6461 1. Create a vexxhost account 2. Create an instance like this: ``` $ openstack flavor list | grep v1-standard-4 | 5cf64088-893b-46b5-9bb1-ee020277635d | v1-standard-4 | 4096 | 100 | 0 | 4 | True | $ openstack image list | grep -i container | 90f57210-9354-4a2f-852e-d844237fbbad | Container Linux 1235.9.0 | active | $ jq . user_data.ign { "ignition": { "version": "2.0.0", "config": {} }, "storage": {}, "systemd": { "units": [ { "name": "docker.service", "enable": true, "dropins": [ { "name": "10-overlay.conf", "contents": "[Service]\nEnvironmentFile=/etc/node.env\nEnvironment=\"DOCKER_OPTS=--storage-driver=overlay2\"\n" } ] } ] }, "networkd": {}, "passwd": {} } $ nova boot \ --image 90f57210-9354-4a2f-852e-d844237fbbad \ --flavor 5cf64088-893b-46b5-9bb1-ee020277635d \ --user-data user_data.ign \ sur-test ``` ### Other Information ### The kernel boot log is available here: https://gist.github.com/s-urbaniak/8108dfa3269da858c489f8cdf38d6461#file-gistfile1-txt /cc @alexsomesan
1.0
Ignition not working on OpenStack - ## Bug ## ### Container Linux Version ### ``` core@localhost ~ $ cat /etc/os-release NAME="Container Linux by CoreOS" ID=coreos VERSION=1235.9.0 VERSION_ID=1235.9.0 BUILD_ID=2017-02-02-0235 PRETTY_NAME="Container Linux by CoreOS 1235.9.0 (Ladybug)" ANSI_COLOR="38;5;75" HOME_URL="https://coreos.com/" BUG_REPORT_URL="https://github.com/coreos/bugs/issues" ``` ### Environment ### Openstack - https://vexxhost.com/ ### Expected Behavior ### Ignition configs should work. ### Actual Behavior ### Ignition configs don't work. ### Reproduction Steps ### https://gist.github.com/s-urbaniak/8108dfa3269da858c489f8cdf38d6461 1. Create a vexxhost account 2. Create an instance like this: ``` $ openstack flavor list | grep v1-standard-4 | 5cf64088-893b-46b5-9bb1-ee020277635d | v1-standard-4 | 4096 | 100 | 0 | 4 | True | $ openstack image list | grep -i container | 90f57210-9354-4a2f-852e-d844237fbbad | Container Linux 1235.9.0 | active | $ jq . user_data.ign { "ignition": { "version": "2.0.0", "config": {} }, "storage": {}, "systemd": { "units": [ { "name": "docker.service", "enable": true, "dropins": [ { "name": "10-overlay.conf", "contents": "[Service]\nEnvironmentFile=/etc/node.env\nEnvironment=\"DOCKER_OPTS=--storage-driver=overlay2\"\n" } ] } ] }, "networkd": {}, "passwd": {} } $ nova boot \ --image 90f57210-9354-4a2f-852e-d844237fbbad \ --flavor 5cf64088-893b-46b5-9bb1-ee020277635d \ --user-data user_data.ign \ sur-test ``` ### Other Information ### The kernel boot log is available here: https://gist.github.com/s-urbaniak/8108dfa3269da858c489f8cdf38d6461#file-gistfile1-txt /cc @alexsomesan
non_process
ignition not working on openstack bug container linux version core localhost cat etc os release name container linux by coreos id coreos version version id build id pretty name container linux by coreos ladybug ansi color home url bug report url environment openstack expected behavior ignition configs should work actual behavior ignition configs don t work reproduction steps create a vexxhost account create an instance like this openstack flavor list grep standard standard true openstack image list grep i container container linux active jq user data ign ignition version config storage systemd units name docker service enable true dropins name overlay conf contents nenvironmentfile etc node env nenvironment docker opts storage driver n networkd passwd nova boot image flavor user data user data ign sur test other information the kernel boot log is available here cc alexsomesan
0
160,852
12,520,775,619
IssuesEvent
2020-06-03 16:23:31
aliasrobotics/RVD
https://api.github.com/repos/aliasrobotics/RVD
opened
Probable insecure usage of temp file/directory, ./src/image_pipeline/camera_calibration/src/camera_calibration/camera_calibrator.py:311
bandit bug components software robot component: ROS static analysis testing triage version: melodic
```yaml { "id": 1, "title": "Probable insecure usage of temp file/directory, ./src/image_pipeline/camera_calibration/src/camera_calibration/camera_calibrator.py:311", "type": "bug", "description": "MEDIUM confidence of MEDIUM severity bug. Probable insecure usage of temp file/directory. ./src/image_pipeline/camera_calibration/src/camera_calibration/camera_calibrator.py:311. See links for more info on the bug.", "cwe": "None", "cve": "None", "keywords": [ "bandit", "bug", "static analysis", "testing", "triage", "bug", "version: melodic", "robot component: ROS", "components software" ], "system": "", "vendor": null, "severity": { "rvss-score": 0, "rvss-vector": "", "severity-description": "", "cvss-score": 0, "cvss-vector": "" }, "links": "", "flaw": { "phase": "testing", "specificity": "subject-specific", "architectural-location": "application-specific", "application": "N/A", "subsystem": "N/A", "package": "N/A", "languages": "None", "date-detected": "2020-06-03 (16:23)", "detected-by": "Alias Robotics", "detected-by-method": "testing static", "date-reported": "2020-06-03 (16:23)", "reported-by": "Alias Robotics", "reported-by-relationship": "automatic", "issue": "", "reproducibility": "always", "trace": "./src/image_pipeline/camera_calibration/src/camera_calibration/camera_calibrator.py:311", "reproduction": "See artifacts below (if available)", "reproduction-image": "" }, "exploitation": { "description": "", "exploitation-image": "", "exploitation-vector": "" }, "mitigation": { "description": "", "pull-request": "", "date-mitigation": "" } } ```
1.0
Probable insecure usage of temp file/directory, ./src/image_pipeline/camera_calibration/src/camera_calibration/camera_calibrator.py:311 - ```yaml { "id": 1, "title": "Probable insecure usage of temp file/directory, ./src/image_pipeline/camera_calibration/src/camera_calibration/camera_calibrator.py:311", "type": "bug", "description": "MEDIUM confidence of MEDIUM severity bug. Probable insecure usage of temp file/directory. ./src/image_pipeline/camera_calibration/src/camera_calibration/camera_calibrator.py:311. See links for more info on the bug.", "cwe": "None", "cve": "None", "keywords": [ "bandit", "bug", "static analysis", "testing", "triage", "bug", "version: melodic", "robot component: ROS", "components software" ], "system": "", "vendor": null, "severity": { "rvss-score": 0, "rvss-vector": "", "severity-description": "", "cvss-score": 0, "cvss-vector": "" }, "links": "", "flaw": { "phase": "testing", "specificity": "subject-specific", "architectural-location": "application-specific", "application": "N/A", "subsystem": "N/A", "package": "N/A", "languages": "None", "date-detected": "2020-06-03 (16:23)", "detected-by": "Alias Robotics", "detected-by-method": "testing static", "date-reported": "2020-06-03 (16:23)", "reported-by": "Alias Robotics", "reported-by-relationship": "automatic", "issue": "", "reproducibility": "always", "trace": "./src/image_pipeline/camera_calibration/src/camera_calibration/camera_calibrator.py:311", "reproduction": "See artifacts below (if available)", "reproduction-image": "" }, "exploitation": { "description": "", "exploitation-image": "", "exploitation-vector": "" }, "mitigation": { "description": "", "pull-request": "", "date-mitigation": "" } } ```
non_process
probable insecure usage of temp file directory src image pipeline camera calibration src camera calibration camera calibrator py yaml id title probable insecure usage of temp file directory src image pipeline camera calibration src camera calibration camera calibrator py type bug description medium confidence of medium severity bug probable insecure usage of temp file directory src image pipeline camera calibration src camera calibration camera calibrator py see links for more info on the bug cwe none cve none keywords bandit bug static analysis testing triage bug version melodic robot component ros components software system vendor null severity rvss score rvss vector severity description cvss score cvss vector links flaw phase testing specificity subject specific architectural location application specific application n a subsystem n a package n a languages none date detected detected by alias robotics detected by method testing static date reported reported by alias robotics reported by relationship automatic issue reproducibility always trace src image pipeline camera calibration src camera calibration camera calibrator py reproduction see artifacts below if available reproduction image exploitation description exploitation image exploitation vector mitigation description pull request date mitigation
0
253,538
21,687,681,332
IssuesEvent
2022-05-09 12:52:12
vector-im/element-ios
https://api.github.com/repos/vector-im/element-ios
opened
Add a close button when leaving spaces
T-Enhancement A-Spaces Z-Community-Testing
### Your use case #### What would you like to do? I think it would be great if we added an X button to the top right corner of the warning that appears when a user tries to leave a space. ![image](https://user-images.githubusercontent.com/18530109/167412692-8836dbba-dc5b-4d34-8761-eedda900d217.png) #### Why would you like to do it? I wanted to cancel leaving a space so I intuitively reached for the top right corner, only to realise that there's nothing there. ### Have you considered any alternatives? The warning can actually be closed by swiping from top to bottom but I feel like there's enough room for a button. ### Additional context _No response_
1.0
Add a close button when leaving spaces - ### Your use case #### What would you like to do? I think it would be great if we added an X button to the top right corner of the warning that appears when a user tries to leave a space. ![image](https://user-images.githubusercontent.com/18530109/167412692-8836dbba-dc5b-4d34-8761-eedda900d217.png) #### Why would you like to do it? I wanted to cancel leaving a space so I intuitively reached for the top right corner, only to realise that there's nothing there. ### Have you considered any alternatives? The warning can actually be closed by swiping from top to bottom but I feel like there's enough room for a button. ### Additional context _No response_
non_process
add a close button when leaving spaces your use case what would you like to do i think it would be great if we added an x button to the top right corner of the warning that appears when a user tries to leave a space why would you like to do it i wanted to cancel leaving a space so i intuitively reached for the top right corner only to realise that there s nothing there have you considered any alternatives the warning can actually be closed by swiping from top to bottom but i feel like there s enough room for a button additional context no response
0