Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
19,541
25,862,260,207
IssuesEvent
2022-12-13 17:49:59
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
DISABLED test_first_argument_index (__main__.SpawnTest)
module: multiprocessing triaged module: flaky-tests skipped
Platforms: linux This test was disabled because it is failing in CI. See [recent examples](http://torch-ci.com/failure/test_first_argument_index%2C%20SpawnTest) and the most recent [workflow logs](https://github.com/pytorch/pytorch/actions/runs/1884385804). Over the past 3 hours, it has been determined flaky in 1 workflow(s) with 1 red and 3 green. cc @VitalyFedyunin
1.0
DISABLED test_first_argument_index (__main__.SpawnTest) - Platforms: linux This test was disabled because it is failing in CI. See [recent examples](http://torch-ci.com/failure/test_first_argument_index%2C%20SpawnTest) and the most recent [workflow logs](https://github.com/pytorch/pytorch/actions/runs/1884385804). Over the past 3 hours, it has been determined flaky in 1 workflow(s) with 1 red and 3 green. cc @VitalyFedyunin
process
disabled test first argument index main spawntest platforms linux this test was disabled because it is failing in ci see and the most recent over the past hours it has been determined flaky in workflow s with red and green cc vitalyfedyunin
1
32,477
13,853,909,949
IssuesEvent
2020-10-15 08:48:15
Azure/azure-sdk-for-net
https://api.github.com/repos/Azure/azure-sdk-for-net
closed
[BUG] IWebAppsOperations.ListSlotConfigurationNamesWithHttpMessagesAsync lacking slotName input
App Services Mgmt Service Attention customer-reported needs-team-attention question
**Describe the bug** The `IWebAppsOperations.ListSlotConfigurationNamesWithHttpMessagesAsync` method lacks an input parameter for the slot name, which causes the method to only work for Web apps (and not Web app **slots**). See also related [issue 12680](https://github.com/Azure/azure-powershell/issues/12680) over at Azure Powershell. **Expected behavior** I can pass a slot name and get the slot config names for the **slot**. **Actual behavior (include Exception or Stack Trace)** The above is not possible. **To Reproduce** Steps to reproduce the behavior (include a code snippet, screenshot, or any additional information that might help us reproduce the issue) 1. Get-AzWebAppSlot to get a slot 2. Use Get-AzWebAppSlotConfigName and pass the resulting object from above to the -WebApp parameter 3. The error "Input object is a deployment slot, not a production web app" is displayed. The error mentioned above might come from the Powershell commandlet, but it throws this because of the bug I'm reporting now. See also [my comment](https://github.com/Azure/azure-powershell/issues/12680#issuecomment-706055722) in the related issue mentioned earlier. **Environment:** - Microsoft.Azure.Management.WebSites (2.2.0, obtained from [this file](https://github.com/Azure/azure-powershell/blob/v4.4.0-July2020/src/Websites/Websites/Websites.csproj), but parameter is lacking in [file on the master branch](https://github.com/Azure/azure-sdk-for-net/blob/master/sdk/websites/Microsoft.Azure.Management.WebSites/src/Generated/IWebAppsOperations.cs)) - See environment data in related issue (mentioned at the top) for more info. - IDE and version : N/A
2.0
[BUG] IWebAppsOperations.ListSlotConfigurationNamesWithHttpMessagesAsync lacking slotName input - **Describe the bug** The `IWebAppsOperations.ListSlotConfigurationNamesWithHttpMessagesAsync` method lacks an input parameter for the slot name, which causes the method to only work for Web apps (and not Web app **slots**). See also related [issue 12680](https://github.com/Azure/azure-powershell/issues/12680) over at Azure Powershell. **Expected behavior** I can pass a slot name and get the slot config names for the **slot**. **Actual behavior (include Exception or Stack Trace)** The above is not possible. **To Reproduce** Steps to reproduce the behavior (include a code snippet, screenshot, or any additional information that might help us reproduce the issue) 1. Get-AzWebAppSlot to get a slot 2. Use Get-AzWebAppSlotConfigName and pass the resulting object from above to the -WebApp parameter 3. The error "Input object is a deployment slot, not a production web app" is displayed. The error mentioned above might come from the Powershell commandlet, but it throws this because of the bug I'm reporting now. See also [my comment](https://github.com/Azure/azure-powershell/issues/12680#issuecomment-706055722) in the related issue mentioned earlier. **Environment:** - Microsoft.Azure.Management.WebSites (2.2.0, obtained from [this file](https://github.com/Azure/azure-powershell/blob/v4.4.0-July2020/src/Websites/Websites/Websites.csproj), but parameter is lacking in [file on the master branch](https://github.com/Azure/azure-sdk-for-net/blob/master/sdk/websites/Microsoft.Azure.Management.WebSites/src/Generated/IWebAppsOperations.cs)) - See environment data in related issue (mentioned at the top) for more info. - IDE and version : N/A
non_process
iwebappsoperations listslotconfigurationnameswithhttpmessagesasync lacking slotname input describe the bug the iwebappsoperations listslotconfigurationnameswithhttpmessagesasync method lacks an input parameter for the slot name which causes the method to only work for web apps and not web app slots see also related over at azure powershell expected behavior i can pass a slot name and get the slot config names for the slot actual behavior include exception or stack trace the above is not possible to reproduce steps to reproduce the behavior include a code snippet screenshot or any additional information that might help us reproduce the issue get azwebappslot to get a slot use get azwebappslotconfigname and pass the resulting object from above to the webapp parameter the error input object is a deployment slot not a production web app is displayed the error mentioned above might come from the powershell commandlet but it throws this because of the bug i m reporting now see also in the related issue mentioned earlier environment microsoft azure management websites obtained from but parameter is lacking in see environment data in related issue mentioned at the top for more info ide and version n a
0
8,296
11,460,852,253
IssuesEvent
2020-02-07 10:35:54
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
Obsolete GO:0044607 disruption by symbiont of host endothelial cells
multi-species process
NO annotations NO reference Should be annotated with an extension.
1.0
Obsolete GO:0044607 disruption by symbiont of host endothelial cells - NO annotations NO reference Should be annotated with an extension.
process
obsolete go disruption by symbiont of host endothelial cells no annotations no reference should be annotated with an extension
1
2,799
5,729,576,746
IssuesEvent
2017-04-21 06:44:41
nodejs/node
https://api.github.com/repos/nodejs/node
opened
doc: typo in doc/api/process.md
doc good first contribution process
* **Version**: master, v7.x, v6.x * **Platform**: n/a * **Subsystem**: doc In doc/api/process.md, there is a typo in the 'A note on process I/O' section where it says: <blockquote> Synchronous writes avoid problems such as output written with `console.log()` or `console.write()` being unexpectedly interleaved, ... </blockquote> where `console.write()` should presumably instead be `console.error()`, since `console.write()` does not exist and `console.error()` is the only other `console` method referred to in that section.
1.0
doc: typo in doc/api/process.md - * **Version**: master, v7.x, v6.x * **Platform**: n/a * **Subsystem**: doc In doc/api/process.md, there is a typo in the 'A note on process I/O' section where it says: <blockquote> Synchronous writes avoid problems such as output written with `console.log()` or `console.write()` being unexpectedly interleaved, ... </blockquote> where `console.write()` should presumably instead be `console.error()`, since `console.write()` does not exist and `console.error()` is the only other `console` method referred to in that section.
process
doc typo in doc api process md version master x x platform n a subsystem doc in doc api process md there is a typo in the a note on process i o section where it says synchronous writes avoid problems such as output written with console log or console write being unexpectedly interleaved where console write should presumably instead be console error since console write does not exist and console error is the only other console method referred to in that section
1
20,386
27,044,891,709
IssuesEvent
2023-02-13 09:04:20
camunda/issues
https://api.github.com/repos/camunda/issues
opened
Decision evaluation outside of Processes
component:operate component:zeebe component:zeebe-process-automation public feature-parity target:8.2
> This is an epic internal-docs issue. It bundles all activities we conduct around a certain initiative. It will typically links to various child issues from various repositories and can be spread across multiple teams. ### Value Proposition Statement Evaluate DMN Decision Tables without a BPMN process by making use of any client. ### User Problem - Currently it is not possible for users to use DMN without a process (e.g. just evaluating a decision). - Users have to build another process, just to evaluate decisions. - This means: a) additional effort when building and deploying DMNs, but more importantly b) leads to high additional costs for the customer (because for each decision evaluation, a process instance is created). c) leads to additional overhead in the engine which can theoretically have negative impact on throughput of evaluations ### User Stories - As a developer, I can evaluate a DMN by using the Java Zeebe Client and see the result as response. - As a developer, I can evaluate a DMN by using the Go Zeebe Client and see the result as response. - As a developer, I can see the result of this evaluation in Operate. Nice-To-have - As a developer, I can evaluate a DMN by using zbctl and see the result as response. - As a developer, I can evaluate a DMN by using the JavaScript client and see the result in Operate. With this Epic we are not focusing on the experience in Modelers, but instead focus on APIs and Clients themselves. There is a separate Epic around making Deployment and Usage of DMNs easier in Modelers. This should not be part of this Epic. ### Implementation Notes ### Links - [Support case](https://jira.camunda.com/browse/SUPPORT-14490) - [Slack channel](https://join.slack.com/share/enQtNDQwODI0NDY1NTY5OC0xOTQ2YWI0NzFjZDI5ODYzZmMzZTMzNmM0NDJkZDAzZTg3ODJiODBhYjE5M2I3ZmVlOWZkOWFiNmE1ZjIzMWVh) #### Design Tickets #### Design ## Design Planning * Reviewed by design: May 17. 2022 * Designer assigned: No Design Necessary #### Implementation Tickets - Zeebe: - https://github.com/camunda/zeebe/issues/11037 - Operate: - https://github.com/camunda/operate/issues/3949 - Optimize: - https://github.com/camunda/camunda-optimize/issues/5806 #### Doc Tickets - https://github.com/camunda/camunda-platform-docs/issues/1516 - https://github.com/camunda/camunda-platform-docs/issues/1519 (stretch goal)
1.0
Decision evaluation outside of Processes - > This is an epic internal-docs issue. It bundles all activities we conduct around a certain initiative. It will typically links to various child issues from various repositories and can be spread across multiple teams. ### Value Proposition Statement Evaluate DMN Decision Tables without a BPMN process by making use of any client. ### User Problem - Currently it is not possible for users to use DMN without a process (e.g. just evaluating a decision). - Users have to build another process, just to evaluate decisions. - This means: a) additional effort when building and deploying DMNs, but more importantly b) leads to high additional costs for the customer (because for each decision evaluation, a process instance is created). c) leads to additional overhead in the engine which can theoretically have negative impact on throughput of evaluations ### User Stories - As a developer, I can evaluate a DMN by using the Java Zeebe Client and see the result as response. - As a developer, I can evaluate a DMN by using the Go Zeebe Client and see the result as response. - As a developer, I can see the result of this evaluation in Operate. Nice-To-have - As a developer, I can evaluate a DMN by using zbctl and see the result as response. - As a developer, I can evaluate a DMN by using the JavaScript client and see the result in Operate. With this Epic we are not focusing on the experience in Modelers, but instead focus on APIs and Clients themselves. There is a separate Epic around making Deployment and Usage of DMNs easier in Modelers. This should not be part of this Epic. ### Implementation Notes ### Links - [Support case](https://jira.camunda.com/browse/SUPPORT-14490) - [Slack channel](https://join.slack.com/share/enQtNDQwODI0NDY1NTY5OC0xOTQ2YWI0NzFjZDI5ODYzZmMzZTMzNmM0NDJkZDAzZTg3ODJiODBhYjE5M2I3ZmVlOWZkOWFiNmE1ZjIzMWVh) #### Design Tickets #### Design ## Design Planning * Reviewed by design: May 17. 2022 * Designer assigned: No Design Necessary #### Implementation Tickets - Zeebe: - https://github.com/camunda/zeebe/issues/11037 - Operate: - https://github.com/camunda/operate/issues/3949 - Optimize: - https://github.com/camunda/camunda-optimize/issues/5806 #### Doc Tickets - https://github.com/camunda/camunda-platform-docs/issues/1516 - https://github.com/camunda/camunda-platform-docs/issues/1519 (stretch goal)
process
decision evaluation outside of processes this is an epic internal docs issue it bundles all activities we conduct around a certain initiative it will typically links to various child issues from various repositories and can be spread across multiple teams value proposition statement evaluate dmn decision tables without a bpmn process by making use of any client user problem currently it is not possible for users to use dmn without a process e g just evaluating a decision users have to build another process just to evaluate decisions this means a additional effort when building and deploying dmns but more importantly b leads to high additional costs for the customer because for each decision evaluation a process instance is created c leads to additional overhead in the engine which can theoretically have negative impact on throughput of evaluations user stories as a developer i can evaluate a dmn by using the java zeebe client and see the result as response as a developer i can evaluate a dmn by using the go zeebe client and see the result as response as a developer i can see the result of this evaluation in operate nice to have as a developer i can evaluate a dmn by using zbctl and see the result as response as a developer i can evaluate a dmn by using the javascript client and see the result in operate with this epic we are not focusing on the experience in modelers but instead focus on apis and clients themselves there is a separate epic around making deployment and usage of dmns easier in modelers this should not be part of this epic implementation notes links design tickets design design planning reviewed by design may designer assigned no design necessary implementation tickets zeebe operate optimize doc tickets stretch goal
1
17,089
22,597,390,998
IssuesEvent
2022-06-29 05:31:19
pyanodon/pybugreports
https://api.github.com/repos/pyanodon/pybugreports
closed
Tailings pond pipe connections are offset
confirmed graphics mod:pycoalprocessing
### Mod source PyAE Beta ### Which mod are you having an issue with? - [ ] pyalienlife - [ ] pyalternativeenergy - [X] pycoalprocessing - [ ] pyfusionenergy - [ ] pyhightech - [ ] pyindustry - [ ] pypetroleumhandling - [ ] pypostprocessing - [ ] pyrawores ### Operating system >=Windows 10 ### What kind of issue is this? - [ ] Compatibility - [ ] Locale (names, descriptions, unknown keys) - [X] Graphical - [ ] Crash - [ ] Progression - [ ] Balance - [ ] Pypostprocessing failure - [ ] Other ### What is the problem? The input/output pipes on the tailings pond are offset from where they should be. The pipes float to the side when connected. ### Steps to reproduce _No response_ ### Additional context ![image](https://user-images.githubusercontent.com/65210810/176341329-4e20d877-e752-4b71-9feb-f03dddc1ceab.png) ### Log file _No response_
1.0
Tailings pond pipe connections are offset - ### Mod source PyAE Beta ### Which mod are you having an issue with? - [ ] pyalienlife - [ ] pyalternativeenergy - [X] pycoalprocessing - [ ] pyfusionenergy - [ ] pyhightech - [ ] pyindustry - [ ] pypetroleumhandling - [ ] pypostprocessing - [ ] pyrawores ### Operating system >=Windows 10 ### What kind of issue is this? - [ ] Compatibility - [ ] Locale (names, descriptions, unknown keys) - [X] Graphical - [ ] Crash - [ ] Progression - [ ] Balance - [ ] Pypostprocessing failure - [ ] Other ### What is the problem? The input/output pipes on the tailings pond are offset from where they should be. The pipes float to the side when connected. ### Steps to reproduce _No response_ ### Additional context ![image](https://user-images.githubusercontent.com/65210810/176341329-4e20d877-e752-4b71-9feb-f03dddc1ceab.png) ### Log file _No response_
process
tailings pond pipe connections are offset mod source pyae beta which mod are you having an issue with pyalienlife pyalternativeenergy pycoalprocessing pyfusionenergy pyhightech pyindustry pypetroleumhandling pypostprocessing pyrawores operating system windows what kind of issue is this compatibility locale names descriptions unknown keys graphical crash progression balance pypostprocessing failure other what is the problem the input output pipes on the tailings pond are offset from where they should be the pipes float to the side when connected steps to reproduce no response additional context log file no response
1
10,755
13,543,365,723
IssuesEvent
2020-09-16 18:50:43
department-of-veterans-affairs/notification-api
https://api.github.com/repos/department-of-veterans-affairs/notification-api
closed
Internal DNS entry
Process Task Ready for... Release
**As** a VANotify system owner **I want** VANotify to be accessible on VA network **So that** VA services can communicate with VANotify. **Acceptance criteria:** - [x] dev.api.notifications.va.gov points to our VAEC dev notification api load balancer - [x] staging.api.notifications.va.gov points to our VAEC staging notification api load balancer - [x] api.notifications.va.gov points to our VAEC prod notification api load balancer This story depends on the Infra - need the Load Balancer DSN so that we can create appropriate CNAME records.
1.0
Internal DNS entry - **As** a VANotify system owner **I want** VANotify to be accessible on VA network **So that** VA services can communicate with VANotify. **Acceptance criteria:** - [x] dev.api.notifications.va.gov points to our VAEC dev notification api load balancer - [x] staging.api.notifications.va.gov points to our VAEC staging notification api load balancer - [x] api.notifications.va.gov points to our VAEC prod notification api load balancer This story depends on the Infra - need the Load Balancer DSN so that we can create appropriate CNAME records.
process
internal dns entry as a vanotify system owner i want vanotify to be accessible on va network so that va services can communicate with vanotify acceptance criteria dev api notifications va gov points to our vaec dev notification api load balancer staging api notifications va gov points to our vaec staging notification api load balancer api notifications va gov points to our vaec prod notification api load balancer this story depends on the infra need the load balancer dsn so that we can create appropriate cname records
1
689,911
23,640,030,993
IssuesEvent
2022-08-25 16:12:15
craftercms/craftercms
https://api.github.com/repos/craftercms/craftercms
closed
[studio-ui] Refused to apply style error in console when using tinymce v2
bug priority: low CI validate
### Duplicates - [X] I have searched the existing issues ### Latest version - [ ] The issue is in the latest released 4.0.x - [X] The issue is in the latest released 3.1.x ### Describe the issue We are getting the following error in Console when trying to edit or view a pape that has a Tiny MCE v2. Refused to apply style from 'http://localhost:8080/studio/static-assets/modules/editors/tinymce/v2/tiny_mce/plugins/inlinepopups/skins/cstudio-rte/window.css?version=48fa93ebd08ecd70eec285dfcda197fa3fde2f73' because its MIME type ('text/html') is not a supported stylesheet MIME type, and strict MIME checking is enabled.trying ### Steps to reproduce Steps: 1. Create a site within Crafter (let's say using a Website editorial blueprint) 2. Add a Tiny MCE v2 to the article form definition 3. Open browser inspect > Console tool 4. Right-click and edit or view any article page. 5. When the editor opens, you will see the error in the console ### Relevant log output _No response_ ### Screenshots and/or videos ![2022-07-25_15-59](https://user-images.githubusercontent.com/2996543/180739754-d5f51b6c-bdf0-4306-a188-ac4ad102ff39.png) ![2022-07-25_16-03](https://user-images.githubusercontent.com/2996543/180740141-3125cd05-3762-4974-9564-5b57f07ca8d3.png) ![2022-07-25_16-03_1](https://user-images.githubusercontent.com/2996543/180740156-6f5383d0-84fc-49ad-9cc4-4e4a4cdbd302.png)
1.0
[studio-ui] Refused to apply style error in console when using tinymce v2 - ### Duplicates - [X] I have searched the existing issues ### Latest version - [ ] The issue is in the latest released 4.0.x - [X] The issue is in the latest released 3.1.x ### Describe the issue We are getting the following error in Console when trying to edit or view a pape that has a Tiny MCE v2. Refused to apply style from 'http://localhost:8080/studio/static-assets/modules/editors/tinymce/v2/tiny_mce/plugins/inlinepopups/skins/cstudio-rte/window.css?version=48fa93ebd08ecd70eec285dfcda197fa3fde2f73' because its MIME type ('text/html') is not a supported stylesheet MIME type, and strict MIME checking is enabled.trying ### Steps to reproduce Steps: 1. Create a site within Crafter (let's say using a Website editorial blueprint) 2. Add a Tiny MCE v2 to the article form definition 3. Open browser inspect > Console tool 4. Right-click and edit or view any article page. 5. When the editor opens, you will see the error in the console ### Relevant log output _No response_ ### Screenshots and/or videos ![2022-07-25_15-59](https://user-images.githubusercontent.com/2996543/180739754-d5f51b6c-bdf0-4306-a188-ac4ad102ff39.png) ![2022-07-25_16-03](https://user-images.githubusercontent.com/2996543/180740141-3125cd05-3762-4974-9564-5b57f07ca8d3.png) ![2022-07-25_16-03_1](https://user-images.githubusercontent.com/2996543/180740156-6f5383d0-84fc-49ad-9cc4-4e4a4cdbd302.png)
non_process
refused to apply style error in console when using tinymce duplicates i have searched the existing issues latest version the issue is in the latest released x the issue is in the latest released x describe the issue we are getting the following error in console when trying to edit or view a pape that has a tiny mce refused to apply style from because its mime type text html is not a supported stylesheet mime type and strict mime checking is enabled trying steps to reproduce steps create a site within crafter let s say using a website editorial blueprint add a tiny mce to the article form definition open browser inspect console tool right click and edit or view any article page when the editor opens you will see the error in the console relevant log output no response screenshots and or videos
0
431,751
12,485,005,827
IssuesEvent
2020-05-30 17:30:11
drupal-celebrations/celebrate-drupal-9
https://api.github.com/repos/drupal-celebrations/celebrate-drupal-9
closed
I'm able to add other users' media to my own posts
priority
**Describe the bug** I created a user account, added a video, and then when I went to add media, I found that I could add a video without having to log in. **To Reproduce** Steps to reproduce the behavior: 1. Log in 2. Go to https://celebratedrupal.org/node/add/video 3. Click 'Add Media' 4. See that you can add other users' media **Expected behavior** I would expect to only be able to add my own media, otherwise I could impersonate other users who have uploaded their own. **Screenshots** <img width="1076" alt="Screen Shot 2020-05-28 at 8 28 26 PM" src="https://user-images.githubusercontent.com/481677/83211181-0dc20300-a122-11ea-822c-8683d0163ace.png"> **Additional context** N/A
1.0
I'm able to add other users' media to my own posts - **Describe the bug** I created a user account, added a video, and then when I went to add media, I found that I could add a video without having to log in. **To Reproduce** Steps to reproduce the behavior: 1. Log in 2. Go to https://celebratedrupal.org/node/add/video 3. Click 'Add Media' 4. See that you can add other users' media **Expected behavior** I would expect to only be able to add my own media, otherwise I could impersonate other users who have uploaded their own. **Screenshots** <img width="1076" alt="Screen Shot 2020-05-28 at 8 28 26 PM" src="https://user-images.githubusercontent.com/481677/83211181-0dc20300-a122-11ea-822c-8683d0163ace.png"> **Additional context** N/A
non_process
i m able to add other users media to my own posts describe the bug i created a user account added a video and then when i went to add media i found that i could add a video without having to log in to reproduce steps to reproduce the behavior log in go to click add media see that you can add other users media expected behavior i would expect to only be able to add my own media otherwise i could impersonate other users who have uploaded their own screenshots img width alt screen shot at pm src additional context n a
0
10,537
13,312,186,652
IssuesEvent
2020-08-26 09:22:23
FAIRplus/FAIRification_process
https://api.github.com/repos/FAIRplus/FAIRification_process
closed
An entrance to find checklists for different datatypes
A: FAIRification process stale
__Making sure people can find checklists for their datatype__ A platform similar to DSP might work.
1.0
An entrance to find checklists for different datatypes - __Making sure people can find checklists for their datatype__ A platform similar to DSP might work.
process
an entrance to find checklists for different datatypes making sure people can find checklists for their datatype a platform similar to dsp might work
1
9,653
12,624,942,099
IssuesEvent
2020-06-14 09:13:04
nodejs/node
https://api.github.com/repos/nodejs/node
closed
child_process.spawn() is unreliable on Promise.all() completion
child_process promises
<!-- Thank you for reporting a possible bug in Node.js. Please fill in as much of the template below as you can. Version: output of `node -v` Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows) Subsystem: if known, please specify the affected core module name If possible, please provide code that demonstrates the problem, keeping it as simple and free of external dependencies as you can. --> * **Version**: ``` $ node -v v10.12.0 ``` * **Platform**: ``` $ uname -a Linux ip-172-31-8-12 4.15.0-1021-aws #21-Ubuntu SMP Tue Aug 28 10:23:07 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux ``` * **Subsystem**: ``` ... var spawn = require('child_process').spawn; ... ``` <!-- Please provide more details below this comment. --> * **Setup**: I have a few promises that I fulfill with `Promise.all()`. When fulfilled, within the `finally()` block, I try to spawn a process and detach it. ``` var p1 = new Promise(function(resolve, reject) { ... }); var p2 = new Promise(function(resolve, reject) { ... }); var p3 = new Promise(function(resolve, reject) { ... }); ... Promise.all([p1, p2, p3, ...]) .then(function(vals) { }) .catch(function(errs) { if (errs) console.log(errs); }) .finally(function() { var myScript = path.join(__dirname, '..', 'bin', 'myScript.js'); var myScriptSpawn = spawn(myScript ['-a', '-b', '-c', ...], { stdio : 'ignore', detached : true, shell : '/usr/bin/node', env : process.env }); myScriptSpawn.unref(); }) ``` * **Problem**: Occasionally the process is spawned, occasionally the process is not spawned. This fails approximately 33% of the time. No warning, exception, or error rises when spawn fails. * **Debugging**: The process `myScript` always works if run manually from the `bash` shell, outside of `node`. The process does not get spawned reliably when attempted within `Promise.all()` within `node`. Moving the function out of the `Promise.all()` fulfillment completion stage seems to work consistently.
1.0
child_process.spawn() is unreliable on Promise.all() completion - <!-- Thank you for reporting a possible bug in Node.js. Please fill in as much of the template below as you can. Version: output of `node -v` Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows) Subsystem: if known, please specify the affected core module name If possible, please provide code that demonstrates the problem, keeping it as simple and free of external dependencies as you can. --> * **Version**: ``` $ node -v v10.12.0 ``` * **Platform**: ``` $ uname -a Linux ip-172-31-8-12 4.15.0-1021-aws #21-Ubuntu SMP Tue Aug 28 10:23:07 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux ``` * **Subsystem**: ``` ... var spawn = require('child_process').spawn; ... ``` <!-- Please provide more details below this comment. --> * **Setup**: I have a few promises that I fulfill with `Promise.all()`. When fulfilled, within the `finally()` block, I try to spawn a process and detach it. ``` var p1 = new Promise(function(resolve, reject) { ... }); var p2 = new Promise(function(resolve, reject) { ... }); var p3 = new Promise(function(resolve, reject) { ... }); ... Promise.all([p1, p2, p3, ...]) .then(function(vals) { }) .catch(function(errs) { if (errs) console.log(errs); }) .finally(function() { var myScript = path.join(__dirname, '..', 'bin', 'myScript.js'); var myScriptSpawn = spawn(myScript ['-a', '-b', '-c', ...], { stdio : 'ignore', detached : true, shell : '/usr/bin/node', env : process.env }); myScriptSpawn.unref(); }) ``` * **Problem**: Occasionally the process is spawned, occasionally the process is not spawned. This fails approximately 33% of the time. No warning, exception, or error rises when spawn fails. * **Debugging**: The process `myScript` always works if run manually from the `bash` shell, outside of `node`. The process does not get spawned reliably when attempted within `Promise.all()` within `node`. Moving the function out of the `Promise.all()` fulfillment completion stage seems to work consistently.
process
child process spawn is unreliable on promise all completion thank you for reporting a possible bug in node js please fill in as much of the template below as you can version output of node v platform output of uname a unix or version and or bit windows subsystem if known please specify the affected core module name if possible please provide code that demonstrates the problem keeping it as simple and free of external dependencies as you can version node v platform uname a linux ip aws ubuntu smp tue aug utc gnu linux subsystem var spawn require child process spawn setup i have a few promises that i fulfill with promise all when fulfilled within the finally block i try to spawn a process and detach it var new promise function resolve reject var new promise function resolve reject var new promise function resolve reject promise all then function vals catch function errs if errs console log errs finally function var myscript path join dirname bin myscript js var myscriptspawn spawn myscript stdio ignore detached true shell usr bin node env process env myscriptspawn unref problem occasionally the process is spawned occasionally the process is not spawned this fails approximately of the time no warning exception or error rises when spawn fails debugging the process myscript always works if run manually from the bash shell outside of node the process does not get spawned reliably when attempted within promise all within node moving the function out of the promise all fulfillment completion stage seems to work consistently
1
1,906
4,733,071,050
IssuesEvent
2016-10-19 09:56:14
nodejs/node
https://api.github.com/repos/nodejs/node
closed
feature request - callbacks in exit handlers - to block even if async code is used
feature request process
I'm on the latest version of Node (6.7) I have been using Node for awhile and one thing that seems to be a missing feature is shutdown hooks that are "blocking", even if they feature async code. To "block" in this case would require callbacks. A simple use-case is a child_process that needs to use `process.send()` (with a request/reply) pattern with a parent process. process.send() is async, and so we can't be guaranteed the parent process will receive the message before the parent may exit. for example what we have now is this ('beforeExit' might be deprecated or gone by now): ``` process.on('beforeExit', function () { }); process.on('exit', function (code, signal) { }); ``` what I am looking for is this type of functionality: ``` process.on('beforeExit', function (cb) { setTimeout(cb, 3000); }); process.on('exit', function (err, code, signal) { // if 'beforeExit' handler is in place, then this will only be called when the callback fires // in the beforeExit handler }); ``` does this functionality exist? The temporary solution is to simply put blocking code in these handlers, but that's not always idea. Hope this makes sense :) I believe it's possible to do this just wondering if it is a sensible request. The use case is I have programs which need to run shutdown hooks even if there is a fatal exception in the code, these shutdown hooks have asynchronous code by nature (process.send, for one). So it's a feature that I really need.
1.0
feature request - callbacks in exit handlers - to block even if async code is used - I'm on the latest version of Node (6.7) I have been using Node for awhile and one thing that seems to be a missing feature is shutdown hooks that are "blocking", even if they feature async code. To "block" in this case would require callbacks. A simple use-case is a child_process that needs to use `process.send()` (with a request/reply) pattern with a parent process. process.send() is async, and so we can't be guaranteed the parent process will receive the message before the parent may exit. for example what we have now is this ('beforeExit' might be deprecated or gone by now): ``` process.on('beforeExit', function () { }); process.on('exit', function (code, signal) { }); ``` what I am looking for is this type of functionality: ``` process.on('beforeExit', function (cb) { setTimeout(cb, 3000); }); process.on('exit', function (err, code, signal) { // if 'beforeExit' handler is in place, then this will only be called when the callback fires // in the beforeExit handler }); ``` does this functionality exist? The temporary solution is to simply put blocking code in these handlers, but that's not always idea. Hope this makes sense :) I believe it's possible to do this just wondering if it is a sensible request. The use case is I have programs which need to run shutdown hooks even if there is a fatal exception in the code, these shutdown hooks have asynchronous code by nature (process.send, for one). So it's a feature that I really need.
process
feature request callbacks in exit handlers to block even if async code is used i m on the latest version of node i have been using node for awhile and one thing that seems to be a missing feature is shutdown hooks that are blocking even if they feature async code to block in this case would require callbacks a simple use case is a child process that needs to use process send with a request reply pattern with a parent process process send is async and so we can t be guaranteed the parent process will receive the message before the parent may exit for example what we have now is this beforeexit might be deprecated or gone by now process on beforeexit function process on exit function code signal what i am looking for is this type of functionality process on beforeexit function cb settimeout cb process on exit function err code signal if beforeexit handler is in place then this will only be called when the callback fires in the beforeexit handler does this functionality exist the temporary solution is to simply put blocking code in these handlers but that s not always idea hope this makes sense i believe it s possible to do this just wondering if it is a sensible request the use case is i have programs which need to run shutdown hooks even if there is a fatal exception in the code these shutdown hooks have asynchronous code by nature process send for one so it s a feature that i really need
1
13,159
15,589,253,333
IssuesEvent
2021-03-18 07:46:30
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
opened
APNs - Upgrade to newer - HTTP/2-based Apple Push Notification service (APNs)
Blocker Process: Enhancement
APNs will no longer support the legacy binary protocol after March 31, 2021. Upgrade to newer - HTTP/2-based Apple Push Notification service (APNs) has to be done and tested before the deadline time. Reference - https://developer.apple.com/news/?id=c88acm2b
1.0
APNs - Upgrade to newer - HTTP/2-based Apple Push Notification service (APNs) - APNs will no longer support the legacy binary protocol after March 31, 2021. Upgrade to newer - HTTP/2-based Apple Push Notification service (APNs) has to be done and tested before the deadline time. Reference - https://developer.apple.com/news/?id=c88acm2b
process
apns upgrade to newer http based apple push notification service apns apns will no longer support the legacy binary protocol after march upgrade to newer http based apple push notification service apns has to be done and tested before the deadline time reference
1
8,356
22,146,249,432
IssuesEvent
2022-06-03 12:22:47
Of-Ash-And-Blight/OAAB-Data
https://api.github.com/repos/Of-Ash-And-Blight/OAAB-Data
closed
Dwemer scaffold wall support
dwemer architecture
Piece to allow dwemer scaffolds to be place on a wall without needing vertical supports beneath.
1.0
Dwemer scaffold wall support - Piece to allow dwemer scaffolds to be place on a wall without needing vertical supports beneath.
non_process
dwemer scaffold wall support piece to allow dwemer scaffolds to be place on a wall without needing vertical supports beneath
0
22,581
31,809,703,108
IssuesEvent
2023-09-13 15:59:03
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
MP GO:0006491 N-glycan processing
PomBase missing parentage protein processing and quality control
GO:0006491 N-glycan processing 2017-01-27 | Deleted | RELATION | part of GO:0006487 (protein N-linked glycosylation) I wondered why this is. Now some of the "N-linked glycosylation pathway" no longer slim to glycosylation?
1.0
MP GO:0006491 N-glycan processing - GO:0006491 N-glycan processing 2017-01-27 | Deleted | RELATION | part of GO:0006487 (protein N-linked glycosylation) I wondered why this is. Now some of the "N-linked glycosylation pathway" no longer slim to glycosylation?
process
mp go n glycan processing go n glycan processing deleted relation part of go protein n linked glycosylation i wondered why this is now some of the n linked glycosylation pathway no longer slim to glycosylation
1
22,547
31,723,427,061
IssuesEvent
2023-09-10 17:16:36
TUM-Dev/NavigaTUM
https://api.github.com/repos/TUM-Dev/NavigaTUM
closed
[Bug] Suche zeigt ab zweitem Punkt keine auto complete an
bug search webform delete-after-processing priority: high
Wenn nach 2906.05.016 gesucht wird, wird ab 2906.05. kein autocomplete mehr angezeigt
1.0
[Bug] Suche zeigt ab zweitem Punkt keine auto complete an - Wenn nach 2906.05.016 gesucht wird, wird ab 2906.05. kein autocomplete mehr angezeigt
process
suche zeigt ab zweitem punkt keine auto complete an wenn nach gesucht wird wird ab kein autocomplete mehr angezeigt
1
8,328
11,490,094,592
IssuesEvent
2020-02-11 16:31:24
ESMValGroup/ESMValCore
https://api.github.com/repos/ESMValGroup/ESMValCore
closed
recipe: using exp: [historical, ssp126] with overlapping data
preprocessor
Theoretically it is possible to combine experiments in the recipe as `{exp: [historical, ssp585], ...}`. Practically this does not work as many models have an overlap (e.g. historical going further than 2014) and then this technique fails (see e.g this [recipe](https://github.com/ESMValGroup/ESMValTool/blob/version2_development/esmvaltool/recipes/recipe_cox18nature.yml#L51-L52)). Is there a way to achieve this? Can a per-experiment `end_year` or `start_year` be defined?
1.0
recipe: using exp: [historical, ssp126] with overlapping data - Theoretically it is possible to combine experiments in the recipe as `{exp: [historical, ssp585], ...}`. Practically this does not work as many models have an overlap (e.g. historical going further than 2014) and then this technique fails (see e.g this [recipe](https://github.com/ESMValGroup/ESMValTool/blob/version2_development/esmvaltool/recipes/recipe_cox18nature.yml#L51-L52)). Is there a way to achieve this? Can a per-experiment `end_year` or `start_year` be defined?
process
recipe using exp with overlapping data theoretically it is possible to combine experiments in the recipe as exp practically this does not work as many models have an overlap e g historical going further than and then this technique fails see e g this is there a way to achieve this can a per experiment end year or start year be defined
1
311,727
26,809,212,098
IssuesEvent
2023-02-01 20:50:18
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
Failing test: X-Pack Case API Integration Tests.x-pack/test/cases_api_integration/security_and_spaces/tests/common/cases/migrations·ts - cases security and spaces enabled: basic Common migrations migrations 7.10.0 -> latest stack version "before all" hook for "migrates cases connector"
failed-test Team:ResponseOps Feature:Cases
A test failed on a tracked branch ``` Error: Timeout of 360000ms exceeded. For async tests and hooks, ensure "done()" is called; if returning a Promise, ensure it resolves. (/var/lib/buildkite-agent/builds/kb-n2-4-spot-2797d6f3f1d89dce/elastic/kibana-on-merge/kibana/x-pack/test/cases_api_integration/security_and_spaces/tests/common/cases/migrations.ts) at listOnTimeout (node:internal/timers:559:17) at processTimers (node:internal/timers:502:7) { code: 'ERR_MOCHA_TIMEOUT', timeout: 360000, file: '/var/lib/buildkite-agent/builds/kb-n2-4-spot-2797d6f3f1d89dce/elastic/kibana-on-merge/kibana/x-pack/test/cases_api_integration/security_and_spaces/tests/common/cases/migrations.ts' } ``` First failure: [CI Build - main](https://buildkite.com/elastic/kibana-on-merge/builds/20393#0182f1f8-8b13-4311-949d-f798c1a03a84) <!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack Case API Integration Tests.x-pack/test/cases_api_integration/security_and_spaces/tests/common/cases/migrations·ts","test.name":"cases security and spaces enabled: basic Common migrations migrations 7.10.0 -> latest stack version \"before all\" hook for \"migrates cases connector\"","test.failCount":2}} -->
1.0
Failing test: X-Pack Case API Integration Tests.x-pack/test/cases_api_integration/security_and_spaces/tests/common/cases/migrations·ts - cases security and spaces enabled: basic Common migrations migrations 7.10.0 -> latest stack version "before all" hook for "migrates cases connector" - A test failed on a tracked branch ``` Error: Timeout of 360000ms exceeded. For async tests and hooks, ensure "done()" is called; if returning a Promise, ensure it resolves. (/var/lib/buildkite-agent/builds/kb-n2-4-spot-2797d6f3f1d89dce/elastic/kibana-on-merge/kibana/x-pack/test/cases_api_integration/security_and_spaces/tests/common/cases/migrations.ts) at listOnTimeout (node:internal/timers:559:17) at processTimers (node:internal/timers:502:7) { code: 'ERR_MOCHA_TIMEOUT', timeout: 360000, file: '/var/lib/buildkite-agent/builds/kb-n2-4-spot-2797d6f3f1d89dce/elastic/kibana-on-merge/kibana/x-pack/test/cases_api_integration/security_and_spaces/tests/common/cases/migrations.ts' } ``` First failure: [CI Build - main](https://buildkite.com/elastic/kibana-on-merge/builds/20393#0182f1f8-8b13-4311-949d-f798c1a03a84) <!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack Case API Integration Tests.x-pack/test/cases_api_integration/security_and_spaces/tests/common/cases/migrations·ts","test.name":"cases security and spaces enabled: basic Common migrations migrations 7.10.0 -> latest stack version \"before all\" hook for \"migrates cases connector\"","test.failCount":2}} -->
non_process
failing test x pack case api integration tests x pack test cases api integration security and spaces tests common cases migrations·ts cases security and spaces enabled basic common migrations migrations latest stack version before all hook for migrates cases connector a test failed on a tracked branch error timeout of exceeded for async tests and hooks ensure done is called if returning a promise ensure it resolves var lib buildkite agent builds kb spot elastic kibana on merge kibana x pack test cases api integration security and spaces tests common cases migrations ts at listontimeout node internal timers at processtimers node internal timers code err mocha timeout timeout file var lib buildkite agent builds kb spot elastic kibana on merge kibana x pack test cases api integration security and spaces tests common cases migrations ts first failure latest stack version before all hook for migrates cases connector test failcount
0
96,992
12,195,977,563
IssuesEvent
2020-04-29 18:17:14
Princeton-CDH/mep-django
https://api.github.com/repos/Princeton-CDH/mep-django
opened
circulation activity table mobile design: edge cases
design
the mobile design for the circulation activity table is difficult to achieve because of some of the constraints on how we can use the elements in it (see discussion below). it would be useful to know how important these choices are to the overall look/message of the page: - combining the two dates into one line with a dash separating them - having the edition info floating on the right side >I tried the same technique used on borrowing activities, which absolute-positions content on the right side, but that doesn't work for long edition titles: ><img width="366" alt="Screen Shot 2020-04-28 at 12 02 25 PM" src="https://user-images.githubusercontent.com/4924494/80510103-449cd180-8948-11ea-9acc-802a58b4bbf2.png"> >the main issue is that the layout implies columns or other structure that wraps parts of the row, but `<div>` aren't allowed inside `<tr>`, so we can't use them to break up the `<td>`s into groups that can then be sized accordingly. >it gets even more complicated because we have to give the appearance of collapsing start and end dates into one dash-separated field, but that field usually needs to take a whole line even though either one of its "parts" could be empty. >i think we may need to compromise on some of these designs - the need to represent them semantically as a table make it quite difficult to achieve other complex, non-tabular representations. _Originally posted by @thatbudakguy in https://github.com/Princeton-CDH/mep-django/pull/572#issuecomment-620703693_
1.0
circulation activity table mobile design: edge cases - the mobile design for the circulation activity table is difficult to achieve because of some of the constraints on how we can use the elements in it (see discussion below). it would be useful to know how important these choices are to the overall look/message of the page: - combining the two dates into one line with a dash separating them - having the edition info floating on the right side >I tried the same technique used on borrowing activities, which absolute-positions content on the right side, but that doesn't work for long edition titles: ><img width="366" alt="Screen Shot 2020-04-28 at 12 02 25 PM" src="https://user-images.githubusercontent.com/4924494/80510103-449cd180-8948-11ea-9acc-802a58b4bbf2.png"> >the main issue is that the layout implies columns or other structure that wraps parts of the row, but `<div>` aren't allowed inside `<tr>`, so we can't use them to break up the `<td>`s into groups that can then be sized accordingly. >it gets even more complicated because we have to give the appearance of collapsing start and end dates into one dash-separated field, but that field usually needs to take a whole line even though either one of its "parts" could be empty. >i think we may need to compromise on some of these designs - the need to represent them semantically as a table make it quite difficult to achieve other complex, non-tabular representations. _Originally posted by @thatbudakguy in https://github.com/Princeton-CDH/mep-django/pull/572#issuecomment-620703693_
non_process
circulation activity table mobile design edge cases the mobile design for the circulation activity table is difficult to achieve because of some of the constraints on how we can use the elements in it see discussion below it would be useful to know how important these choices are to the overall look message of the page combining the two dates into one line with a dash separating them having the edition info floating on the right side i tried the same technique used on borrowing activities which absolute positions content on the right side but that doesn t work for long edition titles img width alt screen shot at pm src the main issue is that the layout implies columns or other structure that wraps parts of the row but aren t allowed inside so we can t use them to break up the s into groups that can then be sized accordingly it gets even more complicated because we have to give the appearance of collapsing start and end dates into one dash separated field but that field usually needs to take a whole line even though either one of its parts could be empty i think we may need to compromise on some of these designs the need to represent them semantically as a table make it quite difficult to achieve other complex non tabular representations originally posted by thatbudakguy in
0
2,121
2,665,630,036
IssuesEvent
2015-03-20 21:53:23
learningequality/ka-lite
https://api.github.com/repos/learningequality/ka-lite
closed
Localization documentation refers to path that does not exist
documentation ux issue
Under: https://github.com/learningequality/ka-lite/blob/master/docs/LOCALIZATION-HACKING.md Section 4.1: KA Lite loads in video content using the content/ directory and the json file **kalite/static/data/topics.json** Take a look at that json data. However, there is no data directory: https://github.com/learningequality/ka-lite/tree/master/kalite/static
1.0
Localization documentation refers to path that does not exist - Under: https://github.com/learningequality/ka-lite/blob/master/docs/LOCALIZATION-HACKING.md Section 4.1: KA Lite loads in video content using the content/ directory and the json file **kalite/static/data/topics.json** Take a look at that json data. However, there is no data directory: https://github.com/learningequality/ka-lite/tree/master/kalite/static
non_process
localization documentation refers to path that does not exist under section ka lite loads in video content using the content directory and the json file kalite static data topics json take a look at that json data however there is no data directory
0
5,978
8,796,060,116
IssuesEvent
2018-12-23 00:01:45
kerubistan/kerub
https://api.github.com/repos/kerubistan/kerub
opened
get rid of lazy properties in infinispan data files
cleanup component:data processing
The problem is that they are redundant, calculated data based on the properties of the entity. If I annotate the delegates with @delegate:Transient, that removes the lazy field from the serialized format, but will generate NPE after deserialization (like e.g. restart of the server) - since the delegate property is now null. JSON format could probably help.
1.0
get rid of lazy properties in infinispan data files - The problem is that they are redundant, calculated data based on the properties of the entity. If I annotate the delegates with @delegate:Transient, that removes the lazy field from the serialized format, but will generate NPE after deserialization (like e.g. restart of the server) - since the delegate property is now null. JSON format could probably help.
process
get rid of lazy properties in infinispan data files the problem is that they are redundant calculated data based on the properties of the entity if i annotate the delegates with delegate transient that removes the lazy field from the serialized format but will generate npe after deserialization like e g restart of the server since the delegate property is now null json format could probably help
1
13,849
16,612,444,352
IssuesEvent
2021-06-02 13:10:08
trilinos/Trilinos
https://api.github.com/repos/trilinos/Trilinos
closed
Create CODEOWNERS File
CLOSED_DUE_TO_INACTIVITY Framework tasks MARKED_FOR_CLOSURE process improvement
@trilinos/framework ## Expectations <!--- Tell us what you think should happen, how you think things should work, what you would like to see in the documentation, etc. --> It'd be great for us to include a [CODEOWNERS](https://help.github.com/articles/about-codeowners/) file in the Trilinos repo root. Having such a file makes it such that we can assign default people or @trilinos/teams as reviewers when pull requests are created, depending on what files/packages/etc. are modified. ## Current Behavior <!--- Tell us how the current behavior fails to meet your expectations in some way. --> Currently we rely on the one submitting the pull request to assign an appropriate reviewer if they know of one. ## Motivation and Context <!--- How has this expectation failure affected you? What are you trying to accomplish? Why do we need to address this? What does it have to do with anything? Providing context helps us come up with a solution that is most useful in the real world. --> This would provide some backup to ensure that the right people are reviewing the changes coming in. ## Related Issues <!--- If applicable, let us know how this bug is related to any other open issues: --> * Related to #1583
1.0
Create CODEOWNERS File - @trilinos/framework ## Expectations <!--- Tell us what you think should happen, how you think things should work, what you would like to see in the documentation, etc. --> It'd be great for us to include a [CODEOWNERS](https://help.github.com/articles/about-codeowners/) file in the Trilinos repo root. Having such a file makes it such that we can assign default people or @trilinos/teams as reviewers when pull requests are created, depending on what files/packages/etc. are modified. ## Current Behavior <!--- Tell us how the current behavior fails to meet your expectations in some way. --> Currently we rely on the one submitting the pull request to assign an appropriate reviewer if they know of one. ## Motivation and Context <!--- How has this expectation failure affected you? What are you trying to accomplish? Why do we need to address this? What does it have to do with anything? Providing context helps us come up with a solution that is most useful in the real world. --> This would provide some backup to ensure that the right people are reviewing the changes coming in. ## Related Issues <!--- If applicable, let us know how this bug is related to any other open issues: --> * Related to #1583
process
create codeowners file trilinos framework expectations tell us what you think should happen how you think things should work what you would like to see in the documentation etc it d be great for us to include a file in the trilinos repo root having such a file makes it such that we can assign default people or trilinos teams as reviewers when pull requests are created depending on what files packages etc are modified current behavior tell us how the current behavior fails to meet your expectations in some way currently we rely on the one submitting the pull request to assign an appropriate reviewer if they know of one motivation and context how has this expectation failure affected you what are you trying to accomplish why do we need to address this what does it have to do with anything providing context helps us come up with a solution that is most useful in the real world this would provide some backup to ensure that the right people are reviewing the changes coming in related issues if applicable let us know how this bug is related to any other open issues related to
1
6,749
9,879,329,711
IssuesEvent
2019-06-24 09:45:47
sejongresearch/EmojiRecommend
https://api.github.com/repos/sejongresearch/EmojiRecommend
closed
영어 번역 코드
Preprocessing
파파고, 카카오 등은 글자 수 제한이 있어서 파이썬 라이브러리 사용 -> 구글에서 request가 한번에 많이 들어오면 IP를 일시적으로 차단하는 문제 발생 -> 약 3만개정도 되는 트윗을 300개씩 나눔 [파일은 여기로](https://drive.google.com/open?id=1hUPMA9tZl5D-B5xQKKD3_etzRG6GL1Sp) -> 300개 실시 후 sleep으로 30분 쉬고 다시 실행되도록 코드를 작성함 [코드는 여기로](https://colab.research.google.com/drive/1_15pj3bOhsk6MghPrCrLsA9Fa3QxypgY) - [x] 1~100 - [x] 100~110 - [x] 110~120 - [x] 120~130 - [x] 130~140 - [x] 140~150 - [x] 150~160 - [x] 160~170 - [x] 170~180 - [x] 180~190 - [x] 190~200 --- - [x] 200~210 - [x] 210~220 - [x] 220~230 - [x] 230~240 - [x] 240~250 - [x] 250~260 - [x] 260~270 - [x] 270~280 - [x] 280~290 - [x] 290~300 --- - [x] 300~310 - [x] 310~320 - [x] 320~330 - [x] 330~340 - [x] 340~350 - [x] 350~360 - [x] 360~370 - [x] 370~380 - [x] 380~390 - [x] 390~400 --- - [x] 400~410 - [x] 410~420 - [x] 420~430 - [x] 430~440 - [x] 440~450 - [x] 450~460 - [x] 460~470 - [x] 470~480 - [x] 480~490 - [x] 490~500
1.0
영어 번역 코드 - 파파고, 카카오 등은 글자 수 제한이 있어서 파이썬 라이브러리 사용 -> 구글에서 request가 한번에 많이 들어오면 IP를 일시적으로 차단하는 문제 발생 -> 약 3만개정도 되는 트윗을 300개씩 나눔 [파일은 여기로](https://drive.google.com/open?id=1hUPMA9tZl5D-B5xQKKD3_etzRG6GL1Sp) -> 300개 실시 후 sleep으로 30분 쉬고 다시 실행되도록 코드를 작성함 [코드는 여기로](https://colab.research.google.com/drive/1_15pj3bOhsk6MghPrCrLsA9Fa3QxypgY) - [x] 1~100 - [x] 100~110 - [x] 110~120 - [x] 120~130 - [x] 130~140 - [x] 140~150 - [x] 150~160 - [x] 160~170 - [x] 170~180 - [x] 180~190 - [x] 190~200 --- - [x] 200~210 - [x] 210~220 - [x] 220~230 - [x] 230~240 - [x] 240~250 - [x] 250~260 - [x] 260~270 - [x] 270~280 - [x] 280~290 - [x] 290~300 --- - [x] 300~310 - [x] 310~320 - [x] 320~330 - [x] 330~340 - [x] 340~350 - [x] 350~360 - [x] 360~370 - [x] 370~380 - [x] 380~390 - [x] 390~400 --- - [x] 400~410 - [x] 410~420 - [x] 420~430 - [x] 430~440 - [x] 440~450 - [x] 450~460 - [x] 460~470 - [x] 470~480 - [x] 480~490 - [x] 490~500
process
영어 번역 코드 파파고 카카오 등은 글자 수 제한이 있어서 파이썬 라이브러리 사용 구글에서 request가 한번에 많이 들어오면 ip를 일시적으로 차단하는 문제 발생 약 되는 트윗을 나눔 실시 후 sleep으로 쉬고 다시 실행되도록 코드를 작성함
1
12,982
15,355,695,550
IssuesEvent
2021-03-01 11:25:46
GoogleCloudPlatform/dotnet-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/dotnet-docs-samples
opened
Spanner: Tests faling with duplicate column name.
api: spanner priority: p1 type: process
Affecting all Spanner tests since this is happening in the fixture: Error: Google.Cloud.Spanner.Data.SpannerException : Duplicate column name Albums.MarketingBudget. See sample [CI output](https://source.cloud.google.com/results/invocations/00fb02fd-8e97-40af-b459-c95f324ef2dd/targets/github%2Fdotnet-docs-samples%2Fspanner%2Fapi%2FSpanner.Samples.Tests%2FTestResults/tests).
1.0
Spanner: Tests faling with duplicate column name. - Affecting all Spanner tests since this is happening in the fixture: Error: Google.Cloud.Spanner.Data.SpannerException : Duplicate column name Albums.MarketingBudget. See sample [CI output](https://source.cloud.google.com/results/invocations/00fb02fd-8e97-40af-b459-c95f324ef2dd/targets/github%2Fdotnet-docs-samples%2Fspanner%2Fapi%2FSpanner.Samples.Tests%2FTestResults/tests).
process
spanner tests faling with duplicate column name affecting all spanner tests since this is happening in the fixture error google cloud spanner data spannerexception duplicate column name albums marketingbudget see sample
1
12,353
14,885,820,913
IssuesEvent
2021-01-20 16:10:15
e4exp/paper_manager_abstract
https://api.github.com/repos/e4exp/paper_manager_abstract
opened
Learning Transferable Visual Models From Natural Language Supervision
2021 Natural Language Processing Pretraining Vision-Language Zero Shot _read_later
* https://cdn.openai.com/papers/Learning_Transferable_Visual_Models_From_Natural_Language_Supervision.pdf * 2021 最新のコンピュータビジョンシステムは、あらかじめ決められたオブジェクトカテゴリの固定セットを予測するように訓練されています。 このような制限された形での監視は、他の視覚的概念を特定するためにラベル付けされたデータが必要となるため、その汎用性と有用性が制限されています。 画像についての生のテキストから直接学習することは、はるかに広い監視のソースを活用する有望な代替手段である。 我々は、インターネットから収集した4億組の画像(画像、テキスト)のデータセット上で、どのキャプションがどの画像に合うかを予測するという単純な事前学習が、SOTA画像表現をスクラッチから学習するための効率的でスケーラブルな方法であることを実証している。 事前学習の後、自然言語を用いて学習した視覚概念を参照することで(あるいは新しい概念を記述することで)、モデルを下流のタスクにゼロショットで移行させることができる。 このアプローチの性能を、OCR、動画のアクション認識、ジオローカリゼーション、多くの種類の微細なオブジェクト分類などのタスクにまたがる30以上の異なる既存のコンピュータビジョンデータセットでベンチマークを行うことで研究しています。 このモデルは、ほとんどのタスクに非トリビエントに適用され、データセット固有のトレーニングを必要とせずに、完全に教師付きのベースラインと競合することがよくあります。 例えば、我々はオリジナルのResNet-50の精度をImageNetのゼロショット上で一致させていますが、128万個の訓練例を使用する必要はありません。
1.0
Learning Transferable Visual Models From Natural Language Supervision - * https://cdn.openai.com/papers/Learning_Transferable_Visual_Models_From_Natural_Language_Supervision.pdf * 2021 最新のコンピュータビジョンシステムは、あらかじめ決められたオブジェクトカテゴリの固定セットを予測するように訓練されています。 このような制限された形での監視は、他の視覚的概念を特定するためにラベル付けされたデータが必要となるため、その汎用性と有用性が制限されています。 画像についての生のテキストから直接学習することは、はるかに広い監視のソースを活用する有望な代替手段である。 我々は、インターネットから収集した4億組の画像(画像、テキスト)のデータセット上で、どのキャプションがどの画像に合うかを予測するという単純な事前学習が、SOTA画像表現をスクラッチから学習するための効率的でスケーラブルな方法であることを実証している。 事前学習の後、自然言語を用いて学習した視覚概念を参照することで(あるいは新しい概念を記述することで)、モデルを下流のタスクにゼロショットで移行させることができる。 このアプローチの性能を、OCR、動画のアクション認識、ジオローカリゼーション、多くの種類の微細なオブジェクト分類などのタスクにまたがる30以上の異なる既存のコンピュータビジョンデータセットでベンチマークを行うことで研究しています。 このモデルは、ほとんどのタスクに非トリビエントに適用され、データセット固有のトレーニングを必要とせずに、完全に教師付きのベースラインと競合することがよくあります。 例えば、我々はオリジナルのResNet-50の精度をImageNetのゼロショット上で一致させていますが、128万個の訓練例を使用する必要はありません。
process
learning transferable visual models from natural language supervision 最新のコンピュータビジョンシステムは、あらかじめ決められたオブジェクトカテゴリの固定セットを予測するように訓練されています。 このような制限された形での監視は、他の視覚的概念を特定するためにラベル付けされたデータが必要となるため、その汎用性と有用性が制限されています。 画像についての生のテキストから直接学習することは、はるかに広い監視のソースを活用する有望な代替手段である。 我々は、 (画像、テキスト)のデータセット上で、どのキャプションがどの画像に合うかを予測するという単純な事前学習が、sota画像表現をスクラッチから学習するための効率的でスケーラブルな方法であることを実証している。 事前学習の後、自然言語を用いて学習した視覚概念を参照することで(あるいは新しい概念を記述することで)、モデルを下流のタスクにゼロショットで移行させることができる。 このアプローチの性能を、ocr、動画のアクション認識、ジオローカリゼーション、 。 このモデルは、ほとんどのタスクに非トリビエントに適用され、データセット固有のトレーニングを必要とせずに、完全に教師付きのベースラインと競合することがよくあります。 例えば、我々はオリジナルのresnet 、 。
1
372,085
11,008,902,353
IssuesEvent
2019-12-04 11:28:37
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
m.bild.de - see bug description
browser-focus-geckoview engine-gecko priority-important
<!-- @browser: Firefox Mobile 71.0 --> <!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Mobile; rv:71.0) Gecko/71.0 Firefox/71.0 --> <!-- @reported_with: --> <!-- @extra_labels: browser-focus-geckoview --> **URL**: https://m.bild.de/wa/ll/bild-de/privater-modus-unangemeldet-54578900.bildMobile.html **Browser / Version**: Firefox Mobile 71.0 **Operating System**: Android 8.1.0 **Tested Another Browser**: No **Problem type**: Something else **Description**: Page says that AdBlock active. But it's disabled **Steps to Reproduce**: Problem with AdBlock <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> Submitted in the name of `@derbazi` _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
m.bild.de - see bug description - <!-- @browser: Firefox Mobile 71.0 --> <!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Mobile; rv:71.0) Gecko/71.0 Firefox/71.0 --> <!-- @reported_with: --> <!-- @extra_labels: browser-focus-geckoview --> **URL**: https://m.bild.de/wa/ll/bild-de/privater-modus-unangemeldet-54578900.bildMobile.html **Browser / Version**: Firefox Mobile 71.0 **Operating System**: Android 8.1.0 **Tested Another Browser**: No **Problem type**: Something else **Description**: Page says that AdBlock active. But it's disabled **Steps to Reproduce**: Problem with AdBlock <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> Submitted in the name of `@derbazi` _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
m bild de see bug description url browser version firefox mobile operating system android tested another browser no problem type something else description page says that adblock active but it s disabled steps to reproduce problem with adblock browser configuration none submitted in the name of derbazi from with ❤️
0
7,071
10,220,890,639
IssuesEvent
2019-08-15 22:58:16
rtcharity/eahub.org
https://api.github.com/repos/rtcharity/eahub.org
opened
Figure out how we're prioritizing things and using milestones now
Process
Personally, I feel as though I've rather lost track of things.
1.0
Figure out how we're prioritizing things and using milestones now - Personally, I feel as though I've rather lost track of things.
process
figure out how we re prioritizing things and using milestones now personally i feel as though i ve rather lost track of things
1
21,253
28,376,526,791
IssuesEvent
2023-04-12 21:20:03
GoogleCloudPlatform/dotnet-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/dotnet-docs-samples
closed
Windows integration tests are not being run always
type: process priority: p1 samples
There seems to be a problem with how the diff command is written, and in some cases, it's not yielding folders to run tests for.
1.0
Windows integration tests are not being run always - There seems to be a problem with how the diff command is written, and in some cases, it's not yielding folders to run tests for.
process
windows integration tests are not being run always there seems to be a problem with how the diff command is written and in some cases it s not yielding folders to run tests for
1
8,527
11,704,715,323
IssuesEvent
2020-03-07 11:20:42
Jeffail/benthos
https://api.github.com/repos/Jeffail/benthos
opened
Change `threads` field behaviour
enhancement processors v4
Currently the `threads` field of `pipeline` is used to create an explicit number of processing threads. This can be useful as sometimes it's desired to limit the workload of a pipeline to one single logical thread (conditional sleeping), and sometimes it's useful to spawn a much larger number of threads than CPU cores if the workload is mostly IO bound (these are green threads anyway). However, I suspect that for the vast majority of users the best behaviour for them is to have as many processing threads as there are CPU cores. In this case our defaults are failing as one pipeline is almost never going to be the ideal number. I'd like to keep this value as a configured field because that allows you to have different thread counts for isolated streams running in streams mode. Therefore I propose changing to default to `0` and documenting that a value of 0 means Benthos matches the number of logical CPU threads on the running machine. Since this would be a breaking change for some hypothetical configs I'll wait for Benthos v4 to implement this.
1.0
Change `threads` field behaviour - Currently the `threads` field of `pipeline` is used to create an explicit number of processing threads. This can be useful as sometimes it's desired to limit the workload of a pipeline to one single logical thread (conditional sleeping), and sometimes it's useful to spawn a much larger number of threads than CPU cores if the workload is mostly IO bound (these are green threads anyway). However, I suspect that for the vast majority of users the best behaviour for them is to have as many processing threads as there are CPU cores. In this case our defaults are failing as one pipeline is almost never going to be the ideal number. I'd like to keep this value as a configured field because that allows you to have different thread counts for isolated streams running in streams mode. Therefore I propose changing to default to `0` and documenting that a value of 0 means Benthos matches the number of logical CPU threads on the running machine. Since this would be a breaking change for some hypothetical configs I'll wait for Benthos v4 to implement this.
process
change threads field behaviour currently the threads field of pipeline is used to create an explicit number of processing threads this can be useful as sometimes it s desired to limit the workload of a pipeline to one single logical thread conditional sleeping and sometimes it s useful to spawn a much larger number of threads than cpu cores if the workload is mostly io bound these are green threads anyway however i suspect that for the vast majority of users the best behaviour for them is to have as many processing threads as there are cpu cores in this case our defaults are failing as one pipeline is almost never going to be the ideal number i d like to keep this value as a configured field because that allows you to have different thread counts for isolated streams running in streams mode therefore i propose changing to default to and documenting that a value of means benthos matches the number of logical cpu threads on the running machine since this would be a breaking change for some hypothetical configs i ll wait for benthos to implement this
1
14,254
17,189,223,141
IssuesEvent
2021-07-16 08:32:50
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[IOS] Wrong error messages is displayed for the following scenario
Bug P1 Process: Fixed iOS
Steps 1. Signup with a valid email on the signup screen 2. After navigating to the verification screen, kill the app 3. Open the app 4. Navigate to Sign up screen 5. Enter the registered email id ( entered email in step1) 6. enter all the mandatory fields and click on Submit 7. Observe the error message AR: 'Your session is expired' error message is displayed ![image](https://user-images.githubusercontent.com/71445210/113820859-e6ec4880-9798-11eb-984c-91bc2844e5c2.png) ER : ![image](https://user-images.githubusercontent.com/71445210/113820927-01262680-9799-11eb-9060-80c3575606a8.png)
1.0
[IOS] Wrong error messages is displayed for the following scenario - Steps 1. Signup with a valid email on the signup screen 2. After navigating to the verification screen, kill the app 3. Open the app 4. Navigate to Sign up screen 5. Enter the registered email id ( entered email in step1) 6. enter all the mandatory fields and click on Submit 7. Observe the error message AR: 'Your session is expired' error message is displayed ![image](https://user-images.githubusercontent.com/71445210/113820859-e6ec4880-9798-11eb-984c-91bc2844e5c2.png) ER : ![image](https://user-images.githubusercontent.com/71445210/113820927-01262680-9799-11eb-9060-80c3575606a8.png)
process
wrong error messages is displayed for the following scenario steps signup with a valid email on the signup screen after navigating to the verification screen kill the app open the app navigate to sign up screen enter the registered email id entered email in enter all the mandatory fields and click on submit observe the error message ar your session is expired error message is displayed er
1
236,351
19,534,026,095
IssuesEvent
2021-12-31 00:18:08
rizinorg/rizin
https://api.github.com/repos/rizinorg/rizin
closed
Wrong analysis results of Busybox PowerPC
RzAnalysis test-required PPC
### Work environment | Questions | Answers |------------------------------------------------------|-------------------- | OS/arch/bits (mandatory) | - | File format of the file you reverse (mandatory) | ELF | Architecture/bits of the file (mandatory) | PPC 32bit | `rizin -v` full output, **not truncated** (mandatory) | https://github.com/rizinorg/rizin/commit/0e11486f36aa784b471d8d4ecf291a8d962b4c47 ### Expected behavior Auto-detection of the function boundaries and function arguments. ### Actual behavior ``` rizin tests/bins/elf/busybox-powerpc aaa s 0x10002d70 Vp ``` ![image](https://user-images.githubusercontent.com/203261/124909594-369f1200-e01d-11eb-896f-bda96416484b.png) Also note the invalid detection of the function arguments. ### Additional Logs, screenshots, source code, configuration dump, ... See also https://github.com/rizinorg/rizin/pull/1281#issuecomment-876327317
1.0
Wrong analysis results of Busybox PowerPC - ### Work environment | Questions | Answers |------------------------------------------------------|-------------------- | OS/arch/bits (mandatory) | - | File format of the file you reverse (mandatory) | ELF | Architecture/bits of the file (mandatory) | PPC 32bit | `rizin -v` full output, **not truncated** (mandatory) | https://github.com/rizinorg/rizin/commit/0e11486f36aa784b471d8d4ecf291a8d962b4c47 ### Expected behavior Auto-detection of the function boundaries and function arguments. ### Actual behavior ``` rizin tests/bins/elf/busybox-powerpc aaa s 0x10002d70 Vp ``` ![image](https://user-images.githubusercontent.com/203261/124909594-369f1200-e01d-11eb-896f-bda96416484b.png) Also note the invalid detection of the function arguments. ### Additional Logs, screenshots, source code, configuration dump, ... See also https://github.com/rizinorg/rizin/pull/1281#issuecomment-876327317
non_process
wrong analysis results of busybox powerpc work environment questions answers os arch bits mandatory file format of the file you reverse mandatory elf architecture bits of the file mandatory ppc rizin v full output not truncated mandatory expected behavior auto detection of the function boundaries and function arguments actual behavior rizin tests bins elf busybox powerpc aaa s vp also note the invalid detection of the function arguments additional logs screenshots source code configuration dump see also
0
5,588
8,443,453,214
IssuesEvent
2018-10-18 15:37:46
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
Relax domain requirementrs for DITA to XDITA conref
dita standard enhancement preprocess/conref
## Expected Behavior Conref content from a DITA topic to an XDITA topic. ## Actual Behavior Transform does not conref and gives the following error message: "[DOTX012W]: When you conref another topic or an item in another topic, the domains attribute of the target topic must be equal to or a subset of the current topic's domains attribute. Put your target under an appropriate domain. You can see the messages guide for more help." ## Possible Solution Relax domains to allow conref. Ask @robander for his opinion/proposed solution.
1.0
Relax domain requirementrs for DITA to XDITA conref - ## Expected Behavior Conref content from a DITA topic to an XDITA topic. ## Actual Behavior Transform does not conref and gives the following error message: "[DOTX012W]: When you conref another topic or an item in another topic, the domains attribute of the target topic must be equal to or a subset of the current topic's domains attribute. Put your target under an appropriate domain. You can see the messages guide for more help." ## Possible Solution Relax domains to allow conref. Ask @robander for his opinion/proposed solution.
process
relax domain requirementrs for dita to xdita conref expected behavior conref content from a dita topic to an xdita topic actual behavior transform does not conref and gives the following error message when you conref another topic or an item in another topic the domains attribute of the target topic must be equal to or a subset of the current topic s domains attribute put your target under an appropriate domain you can see the messages guide for more help possible solution relax domains to allow conref ask robander for his opinion proposed solution
1
20,767
27,501,246,867
IssuesEvent
2023-03-05 18:14:45
hsmusic/hsmusic-data
https://api.github.com/repos/hsmusic/hsmusic-data
closed
Locate and add available uncompressed PNG artwork
scope: official scope: fandom scope: beyond type: addition type: involved process type: needs investigation
"Available" obviously means stuff that wasn't locked behind a paywall, but as we've discussed, that's basically all publicly visible artwork on any Bandcamp (`_0.png` suffix bcbits URLs), and whatever can be recovered from official releases, since all the official ways to buy their artwork have been taken offline. Ties in with hsmusic/hsmusic-wiki#70 nicely - we should get these in the same update! (Not the one currently being finalized though.) (We can break this issue into a todo checklist if we want to track progress granularly)
1.0
Locate and add available uncompressed PNG artwork - "Available" obviously means stuff that wasn't locked behind a paywall, but as we've discussed, that's basically all publicly visible artwork on any Bandcamp (`_0.png` suffix bcbits URLs), and whatever can be recovered from official releases, since all the official ways to buy their artwork have been taken offline. Ties in with hsmusic/hsmusic-wiki#70 nicely - we should get these in the same update! (Not the one currently being finalized though.) (We can break this issue into a todo checklist if we want to track progress granularly)
process
locate and add available uncompressed png artwork available obviously means stuff that wasn t locked behind a paywall but as we ve discussed that s basically all publicly visible artwork on any bandcamp png suffix bcbits urls and whatever can be recovered from official releases since all the official ways to buy their artwork have been taken offline ties in with hsmusic hsmusic wiki nicely we should get these in the same update not the one currently being finalized though we can break this issue into a todo checklist if we want to track progress granularly
1
85,094
24,508,709,996
IssuesEvent
2022-10-10 19:01:35
PowerShell/PowerShellGet
https://api.github.com/repos/PowerShell/PowerShellGet
closed
Abstract at least common dependencies into an AssemblyLoadContext
Area-Build&Release feature_request
### Summary of the new feature / enhancement PSGet 3.0 should follow the example of PSES and abstract its common dependencies such as NuGet into an `AssemblyLoadContext`, so as not to conflict with other legacy modules which might use those assemblies. Per @SteveL-MSFT: https://twitter.com/Steve_MSFT/status/1542983766901026816?s=20&t=lguXc5lP3bbwtRahi34EUw ### Proposed technical implementation details (optional) _No response_
1.0
Abstract at least common dependencies into an AssemblyLoadContext - ### Summary of the new feature / enhancement PSGet 3.0 should follow the example of PSES and abstract its common dependencies such as NuGet into an `AssemblyLoadContext`, so as not to conflict with other legacy modules which might use those assemblies. Per @SteveL-MSFT: https://twitter.com/Steve_MSFT/status/1542983766901026816?s=20&t=lguXc5lP3bbwtRahi34EUw ### Proposed technical implementation details (optional) _No response_
non_process
abstract at least common dependencies into an assemblyloadcontext summary of the new feature enhancement psget should follow the example of pses and abstract its common dependencies such as nuget into an assemblyloadcontext so as not to conflict with other legacy modules which might use those assemblies per stevel msft proposed technical implementation details optional no response
0
13,462
4,713,110,788
IssuesEvent
2016-10-14 18:56:57
mozilla/addons-frontend
https://api.github.com/repos/mozilla/addons-frontend
opened
Don't use `development` method on webpackIsomorphicToolsPlugin
code quality
``` [webpack-isomorphic-tools] [error] `.development()` method is now deprecated (for server-side instance only, not for webpack plugin instance) and has no effect. Set up a proper `process.env.NODE_ENV` variable instead. ```
1.0
Don't use `development` method on webpackIsomorphicToolsPlugin - ``` [webpack-isomorphic-tools] [error] `.development()` method is now deprecated (for server-side instance only, not for webpack plugin instance) and has no effect. Set up a proper `process.env.NODE_ENV` variable instead. ```
non_process
don t use development method on webpackisomorphictoolsplugin development method is now deprecated for server side instance only not for webpack plugin instance and has no effect set up a proper process env node env variable instead
0
733,770
25,321,540,975
IssuesEvent
2022-11-18 04:40:07
rpitv/glimpse-graphics
https://api.github.com/repos/rpitv/glimpse-graphics
opened
Announcements timer enhancements
enhancement good first issue Priority: LOW
There are a number of nice-to-have additions to announcements which should all be relatively simple: - Default timer settings - I.e., what's the default behavior for what happens to an announcement when it's timer ends? - These should probably persist across reload, but not between users (i.e., store in local storage) - Remove a timer from an announcement - This can currently be accomplished by just setting the time to 0 and telling it to remove the timer, but that's not very clean. - Add a timer to an announcement after it's already been created
1.0
Announcements timer enhancements - There are a number of nice-to-have additions to announcements which should all be relatively simple: - Default timer settings - I.e., what's the default behavior for what happens to an announcement when it's timer ends? - These should probably persist across reload, but not between users (i.e., store in local storage) - Remove a timer from an announcement - This can currently be accomplished by just setting the time to 0 and telling it to remove the timer, but that's not very clean. - Add a timer to an announcement after it's already been created
non_process
announcements timer enhancements there are a number of nice to have additions to announcements which should all be relatively simple default timer settings i e what s the default behavior for what happens to an announcement when it s timer ends these should probably persist across reload but not between users i e store in local storage remove a timer from an announcement this can currently be accomplished by just setting the time to and telling it to remove the timer but that s not very clean add a timer to an announcement after it s already been created
0
18,885
24,825,057,195
IssuesEvent
2022-10-25 19:51:29
maticnetwork/miden
https://api.github.com/repos/maticnetwork/miden
opened
Remove queued_requests from the hasher
enhancement processor v0.4
Currently, the entire hash computation is done when the decoder makes its initialization request and all intermediate lookups required for the correctness of $b_{chip}$ are queued. When the decoder needs subsequent lookups (e.g. as it absorbs new operations during RESPAN or when the completes code blocks and needs the return hash), it sends a request, and they are dequeued and sent to the $b_{chip}$ bus. Instead, it might be better to compute the lookups at the time they are needed. Solution proposed by @bobbinth: The entire trace for a span block would still be computed as a part of the `Hasher::hash_span_block()`. But this method would not create any lookups. Instead, the `HasherLookupContext::Start` lookup would be created in `Chiplets::hash_span_block()`, and all other lookups would be created by `Chiplets::absorb_span_batch()` and `Chiplets::read_hash_result()` (though, we'd probably need to have a specialized version of this method for SPAN blocks). There is a bit more complexity here as `Decoder::respan()` doesn't have access to the address of the batch being executed but I think it shouldn't be too complicated to provide this info to this method, and once it is there, other parts should be relatively straight-forward. So, the `Hasher` would no longer be responsible for computing lookups, and that responsibility would move into hasher-related Chiplets methods. For control blocks, the solution should be relatively simple: 1. At the time when control block execution ends, the decoder knows the address of the block (which it received from the hasher at the time when block hash is computed). We would need to update some method signatures in the decoder to propagate this info back to `end_join_block()`, `end_split_block()` etc. methods. 2. We can then pass this address into `Chiplets::read_hash_result()` method, and that should be enough to compute a lookup for the return value (addr for the lookup would be addr of the block + 7).
1.0
Remove queued_requests from the hasher - Currently, the entire hash computation is done when the decoder makes its initialization request and all intermediate lookups required for the correctness of $b_{chip}$ are queued. When the decoder needs subsequent lookups (e.g. as it absorbs new operations during RESPAN or when the completes code blocks and needs the return hash), it sends a request, and they are dequeued and sent to the $b_{chip}$ bus. Instead, it might be better to compute the lookups at the time they are needed. Solution proposed by @bobbinth: The entire trace for a span block would still be computed as a part of the `Hasher::hash_span_block()`. But this method would not create any lookups. Instead, the `HasherLookupContext::Start` lookup would be created in `Chiplets::hash_span_block()`, and all other lookups would be created by `Chiplets::absorb_span_batch()` and `Chiplets::read_hash_result()` (though, we'd probably need to have a specialized version of this method for SPAN blocks). There is a bit more complexity here as `Decoder::respan()` doesn't have access to the address of the batch being executed but I think it shouldn't be too complicated to provide this info to this method, and once it is there, other parts should be relatively straight-forward. So, the `Hasher` would no longer be responsible for computing lookups, and that responsibility would move into hasher-related Chiplets methods. For control blocks, the solution should be relatively simple: 1. At the time when control block execution ends, the decoder knows the address of the block (which it received from the hasher at the time when block hash is computed). We would need to update some method signatures in the decoder to propagate this info back to `end_join_block()`, `end_split_block()` etc. methods. 2. We can then pass this address into `Chiplets::read_hash_result()` method, and that should be enough to compute a lookup for the return value (addr for the lookup would be addr of the block + 7).
process
remove queued requests from the hasher currently the entire hash computation is done when the decoder makes its initialization request and all intermediate lookups required for the correctness of b chip are queued when the decoder needs subsequent lookups e g as it absorbs new operations during respan or when the completes code blocks and needs the return hash it sends a request and they are dequeued and sent to the b chip bus instead it might be better to compute the lookups at the time they are needed solution proposed by bobbinth the entire trace for a span block would still be computed as a part of the hasher hash span block but this method would not create any lookups instead the hasherlookupcontext start lookup would be created in chiplets hash span block and all other lookups would be created by chiplets absorb span batch and chiplets read hash result though we d probably need to have a specialized version of this method for span blocks there is a bit more complexity here as decoder respan doesn t have access to the address of the batch being executed but i think it shouldn t be too complicated to provide this info to this method and once it is there other parts should be relatively straight forward so the hasher would no longer be responsible for computing lookups and that responsibility would move into hasher related chiplets methods for control blocks the solution should be relatively simple at the time when control block execution ends the decoder knows the address of the block which it received from the hasher at the time when block hash is computed we would need to update some method signatures in the decoder to propagate this info back to end join block end split block etc methods we can then pass this address into chiplets read hash result method and that should be enough to compute a lookup for the return value addr for the lookup would be addr of the block
1
411,641
27,827,065,459
IssuesEvent
2023-03-19 21:54:26
J99thoms/Super-Smash-Dashboard
https://api.github.com/repos/J99thoms/Super-Smash-Dashboard
opened
Ensure Reproducibility and Deploy the App
documentation enhancement
rubric={mechanics: 30} - [ ] Deploy the app using Render and include the link on the GitHub repo where the user can find it easily (`README.md` or about section). - [ ] Add in the README.md a reference section to the dataset you are using.
1.0
Ensure Reproducibility and Deploy the App - rubric={mechanics: 30} - [ ] Deploy the app using Render and include the link on the GitHub repo where the user can find it easily (`README.md` or about section). - [ ] Add in the README.md a reference section to the dataset you are using.
non_process
ensure reproducibility and deploy the app rubric mechanics deploy the app using render and include the link on the github repo where the user can find it easily readme md or about section add in the readme md a reference section to the dataset you are using
0
10,852
13,629,166,837
IssuesEvent
2020-09-24 14:47:33
KantaraInitiative/wg-uma
https://api.github.com/repos/KantaraInitiative/wg-uma
closed
Error response and error response registration issues
V2.0 fedauthz grant process
- After doing a bunch of token endpoint error alignment/rationalization in [UMA2 Core rev 17](https://docs.kantarainitiative.org/uma/ed/uma-core-2.0-17.html), what I came out with was invalid_grant (mentioning ticket problems; expired_ticket and invalid_ticket go away), invalid_scope (redefining it for the UMA grant and using HTTP 400 instead of 401), not_authorized (new), request_submitted (new), and need_info with its contents (new). OAuth's other errors don't change. The only one that's a bit weird to me is redefining invalid_scope so that OAuth's original HTTP 401 error goes to an HTTP 400 error. But I wouldn't want to rename it either, since we add an invalid_scope error from the AS to the RS for the protection API, and the name is perfectly accurate. Should we simply change to producing a 401? - I labeled each of our registry sections with the authoritative section for the template we're following. This led me to question our having a [JWT claims registration](https://docs.kantarainitiative.org/uma/ed/uma-core-2.0-17.html#rfc.section.9.1) (source: OIDC, I think!) along with also having a [token introspection response registration](https://docs.kantarainitiative.org/uma/ed/uma-core-2.0-17.html#rfc.section.9.3). Do we actually need both? - What's the right timing for sending all over our various registration requests to oauth-ext-review@ietf.org and wellknown-uri-review@ietf.org, as the referenced registration templates say they require? Draft stage or final stage?
1.0
Error response and error response registration issues - - After doing a bunch of token endpoint error alignment/rationalization in [UMA2 Core rev 17](https://docs.kantarainitiative.org/uma/ed/uma-core-2.0-17.html), what I came out with was invalid_grant (mentioning ticket problems; expired_ticket and invalid_ticket go away), invalid_scope (redefining it for the UMA grant and using HTTP 400 instead of 401), not_authorized (new), request_submitted (new), and need_info with its contents (new). OAuth's other errors don't change. The only one that's a bit weird to me is redefining invalid_scope so that OAuth's original HTTP 401 error goes to an HTTP 400 error. But I wouldn't want to rename it either, since we add an invalid_scope error from the AS to the RS for the protection API, and the name is perfectly accurate. Should we simply change to producing a 401? - I labeled each of our registry sections with the authoritative section for the template we're following. This led me to question our having a [JWT claims registration](https://docs.kantarainitiative.org/uma/ed/uma-core-2.0-17.html#rfc.section.9.1) (source: OIDC, I think!) along with also having a [token introspection response registration](https://docs.kantarainitiative.org/uma/ed/uma-core-2.0-17.html#rfc.section.9.3). Do we actually need both? - What's the right timing for sending all over our various registration requests to oauth-ext-review@ietf.org and wellknown-uri-review@ietf.org, as the referenced registration templates say they require? Draft stage or final stage?
process
error response and error response registration issues after doing a bunch of token endpoint error alignment rationalization in what i came out with was invalid grant mentioning ticket problems expired ticket and invalid ticket go away invalid scope redefining it for the uma grant and using http instead of not authorized new request submitted new and need info with its contents new oauth s other errors don t change the only one that s a bit weird to me is redefining invalid scope so that oauth s original http error goes to an http error but i wouldn t want to rename it either since we add an invalid scope error from the as to the rs for the protection api and the name is perfectly accurate should we simply change to producing a i labeled each of our registry sections with the authoritative section for the template we re following this led me to question our having a source oidc i think along with also having a do we actually need both what s the right timing for sending all over our various registration requests to oauth ext review ietf org and wellknown uri review ietf org as the referenced registration templates say they require draft stage or final stage
1
92,911
26,802,051,764
IssuesEvent
2023-02-01 15:43:54
flutter/flutter
https://api.github.com/repos/flutter/flutter
closed
flutter build --release crashes 3.7.0
severe: crash tool dependency: dart a: build
<!-- Thank you for using Flutter! If you are looking for support, please check out our documentation or consider asking a question on Stack Overflow: * https://flutter.dev/ * https://api.flutter.dev/ * https://stackoverflow.com/questions/tagged/flutter?sort=frequent If you have found a bug or if our documentation doesn't have an answer to what you're looking for, then fill out the template below. Please read our guide to filing a bug first: https://flutter.dev/docs/resources/bug-reports --> ## Steps to Reproduce <!-- Please tell us exactly how to reproduce the problem you are running into. --> Building a debug app works `flutter build apk --debug --flavor dev -t lib/main.dart` ```sh Running Gradle task 'assembleDevDebug'... 33.5s ✓ Built build/app/outputs/flutter-apk/app-dev-debug.apk. ``` But building for release crashes `flutter build apk --release --flavor dev -t lib/main.dart` ```sh ===== CRASH ===== si_signo=Segmentation fault: 11(11), si_code=1, si_addr=0xc version=2.19.0 (stable) (Mon Jan 23 11:29:09 2023 -0800) on "macos_x64" pid=61474, thread=-1, isolate_group=isolate(0x7fec27808200), isolate=(nil)(0x0) os=macos, arch=x64, comp=yes, sim=no isolate_instructions=0, vm_instructions=0 ``` Before with Flutter 3.3.10 --release build successful. This is a App in released in production for Android and iOS ## Logs <details> <summary>Logs</summary> <!-- Include the full logs of the commands you are running between the lines with the backticks below. If you are running any "flutter" commands, please include the output of running them with "--verbose"; for example, the output of running "flutter --verbose create foo". --> ``` 💪 Building with sound null safety 💪 Expected to find fonts for (packages/cupertino_icons/CupertinoIcons, MaterialIcons, WuerthIcon), but found (MaterialIcons, WuerthIcon). This usually means you are referring to font families in an IconData class but not including them in the assets section of your pubspec.yaml, are missing the package that would include them, or are missing "uses-material-design: true". ===== CRASH ===== si_signo=Segmentation fault: 11(11), si_code=1, si_addr=0xc version=2.19.0 (stable) (Mon Jan 23 11:29:09 2023 -0800) on "macos_arm" pid=59850, thread=-1, isolate_group=isolate(0x7fe1b801b400), isolate=(nil)(0x0) os=macos, arch=arm, comp=no, sim=no isolate_instructions=0, vm_instructions=0 pc 0x000000010139e2ad fp 0x0000000306bfe100 dart::FlowGraphTypePropagator::VisitValue(dart::Value*)+0x3d pc 0x000000010139dee7 fp 0x0000000306bfe150 dart::FlowGraphTypePropagator::PropagateRecursive(dart::BlockEntryInstr*)+0xd7 pc 0x000000010139dfc3 fp 0x0000000306bfe1a0 dart::FlowGraphTypePropagator::PropagateRecursive(dart::BlockEntryInstr*)+0x1b3 pc 0x000000010139d5a2 fp 0x0000000306bfe620 dart::FlowGraphTypePropagator::Propagate()+0x32 pc 0x000000010139d538 fp 0x0000000306bfe740 dart::FlowGraphTypePropagator::Propagate(dart::FlowGraph*)+0x58 pc 0x00000001013b5662 fp 0x0000000306bfe760 dart::CompilerPass_TypePropagation::DoBody(dart::CompilerPassState*) const+0x22 pc 0x00000001013b4250 fp 0x0000000306bfe830 dart::CompilerPass::Run(dart::CompilerPassState*) const+0x1b0 pc 0x00000001013b447a fp 0x0000000306bfe860 dart::CompilerPass::RunInliningPipeline(dart::CompilerPass::PipelineMode, dart::CompilerPassState*)+0x3a pc 0x00000001013630fb fp 0x0000000306bfed70 dart::CallSiteInliner::TryInliningImpl(dart::Function const&, dart::Array const&, dart::InlinedCallData*, bool)+0x100b pc 0x00000001013577d6 fp 0x0000000306bfee10 dart::CallSiteInliner::TryInlining(dart::Function const&, dart::Array const&, dart::InlinedCallData*, bool)+0xf6 pc 0x00000001013662db fp 0x0000000306bfef00 dart::CallSiteInliner::InlineStaticCalls()+0x56b pc 0x000000010135d3df fp 0x0000000306bff020 dart::CallSiteInliner::InlineCalls()+0x24f pc 0x000000010135d031 fp 0x0000000306bff120 dart::FlowGraphInliner::Inline()+0x1e1 pc 0x00000001013b55d2 fp 0x0000000306bff190 dart::CompilerPass_Inlining::DoBody(dart::CompilerPassState*) const+0x52 pc 0x00000001013b4250 fp 0x0000000306bff260 dart::CompilerPass::Run(dart::CompilerPassState*) const+0x1b0 pc 0x00000001013b4739 fp 0x0000000306bff290 dart::CompilerPass::RunPipeline(dart::CompilerPass::PipelineMode, dart::CompilerPassState*)+0xa9 pc 0x0000000101286bf9 fp 0x0000000306bff960 dart::PrecompileParsedFunctionHelper::Compile(dart::CompilationPipeline*)+0x5c9 pc 0x00000001012876fd fp 0x0000000306bfffe0 dart::PrecompileFunctionHelper(dart::Precompiler*, dart::CompilationPipeline*, dart::Function const&, bool)+0x31d pc 0x00000001012824f6 fp 0x0000000306c000f0 dart::Precompiler::CompileFunction(dart::Precompiler*, dart::Thread*, dart::Zone*, dart::Function const&)+0x176 pc 0x000000010127fd81 fp 0x0000000306c00190 dart::Precompiler::ProcessFunction(dart::Function const&)+0xf1 pc 0x000000010127a054 fp 0x0000000306c001f0 dart::Precompiler::Iterate()+0x94 pc 0x0000000101274b19 fp 0x0000000306c00940 dart::Precompiler::DoCompileAll()+0x1489 pc 0x000000010127360b fp 0x0000000306c00da0 dart::Precompiler::CompileAll()+0xbb pc 0x0000000101468ab8 fp 0x0000000306c00f10 Dart_Precompile+0x248 pc 0x0000000100fc2f4b fp 0x0000000306c010a0 dart::bin::main(int, char**)+0x9ab pc 0x0000000201b4f310 fp 0x0000000306c01340 Unknown symbol -- End of DumpStackTrace === Crash occurred when compiling package:photodoc/model/dto/project_dto.dart_ProjectDto_ProjectDto.init in AOT mode in TypePropagation pass *** BEGIN CFG TypePropagation ==== package:photodoc/model/dto/project_dto.dart__$_ProjectDto@1727099614__$_ProjectDto@1727099614. (Constructor) B0[graph]:0 { v0 <- Constant(#-1) T{_Smi} v1 <- Constant(#) T{_OneByteString} v2 <- Constant(#false) T{bool} v3 <- Constant(#null) T{Null?} v4 <- Constant(#_ImmutableList len:0) v5 <- Constant(#_ImmutableList len:0) v6 <- Constant(#_ImmutableList len:0) v7 <- Constant(#_ImmutableList len:0) v8 <- Constant(#true) v5 <- Constant(#<optimized out>) } B3[function entry]:2 { v6 <- Parameter(0) T{_$_ProjectDto} v7 <- SpecialParameter(ArgDescriptor) } StoreField(v6 . id = v0 <int64>) StoreField(v6 . name = v1) StoreField(v6 . eventStoreId = v1) StoreField(v6 . deleted = v2) StoreField(v6 . lastUpdatedByUser = v3) StoreField(v6 . createdDate = v3) StoreField(v6 . createdBy = v3) StoreField(v6 . company = v3) StoreField(v6 . address = v3) StoreField(v6 . customer = v3) StoreField(v6 . distanceToCurrentUser = v1) StoreField(v6 . number = v1) StoreField(v6 . isSynchronizeEnabled = v8) StoreField(v6 . isNewProject = v2) StoreField(v6 . isOffline = v2) StoreField(v6 . _images@1727099614 = v4) StoreField(v6 . _tags@1727099614 = v5) StoreField(v6 . _reports@1727099614 = v6) StoreField(v6 . _participants@1727099614 = v7) Return:470(v3) *** END CFG Dart snapshot generator failed with exit code -6 ===== CRASH ===== si_signo=Segmentation fault: 11(11), si_code=1, si_addr=0xc version=2.19.0 (stable) (Mon Jan 23 11:29:09 2023 -0800) on "macos_simarm64" pid=59851, thread=-1, isolate_group=isolate(0x7fd545008200), isolate=(nil)(0x0) os=macos, arch=arm64, comp=yes, sim=yes isolate_instructions=0, vm_instructions=0 pc 0x000000010490b1cd fp 0x000000030d841140 dart::FlowGraphTypePropagator::VisitValue(dart::Value*)+0x3d pc 0x000000010490ae07 fp 0x000000030d841190 dart::FlowGraphTypePropagator::PropagateRecursive(dart::BlockEntryInstr*)+0xd7 pc 0x000000010490aee3 fp 0x000000030d8411e0 dart::FlowGraphTypePropagator::PropagateRecursive(dart::BlockEntryInstr*)+0x1b3 pc 0x000000010490a4c2 fp 0x000000030d841660 dart::FlowGraphTypePropagator::Propagate()+0x32 pc 0x000000010490a458 fp 0x000000030d841780 dart::FlowGraphTypePropagator::Propagate(dart::FlowGraph*)+0x58 pc 0x0000000104922c12 fp 0x000000030d8417a0 dart::CompilerPass_TypePropagation::DoBody(dart::CompilerPassState*) const+0x22 pc 0x0000000104921800 fp 0x000000030d841870 dart::CompilerPass::Run(dart::CompilerPassState*) const+0x1b0 pc 0x0000000104921a2a fp 0x000000030d8418a0 dart::CompilerPass::RunInliningPipeline(dart::CompilerPass::PipelineMode, dart::CompilerPassState*)+0x3a pc 0x00000001048d0393 fp 0x000000030d841db0 dart::CallSiteInliner::TryInliningImpl(dart::Function const&, dart::Array const&, dart::InlinedCallData*, bool)+0x1073 pc 0x00000001048c48f6 fp 0x000000030d841e50 dart::CallSiteInliner::TryInlining(dart::Function const&, dart::Array const&, dart::InlinedCallData*, bool)+0xf6 pc 0x00000001048d36eb fp 0x000000030d841f40 dart::CallSiteInliner::InlineStaticCalls()+0x56b pc 0x00000001048ca64f fp 0x000000030d842060 dart::CallSiteInliner::InlineCalls()+0x24f pc 0x00000001048ca2a1 fp 0x000000030d842160 dart::FlowGraphInliner::Inline()+0x1e1 pc 0x0000000104922b82 fp 0x000000030d8421d0 dart::CompilerPass_Inlining::DoBody(dart::CompilerPassState*) const+0x52 pc 0x0000000104921800 fp 0x000000030d8422a0 dart::CompilerPass::Run(dart::CompilerPassState*) const+0x1b0 pc 0x0000000104921ce9 fp 0x000000030d8422d0 dart::CompilerPass::RunPipeline(dart::CompilerPass::PipelineMode, dart::CompilerPassState*)+0xa9 pc 0x00000001047f9fa1 fp 0x000000030d8429a0 dart::PrecompileParsedFunctionHelper::Compile(dart::CompilationPipeline*)+0x611 pc 0x00000001047faa9a fp 0x000000030d843020 dart::PrecompileFunctionHelper(dart::Precompiler*, dart::CompilationPipeline*, dart::Function const&, bool)+0x31a pc 0x00000001047f4e56 fp 0x000000030d843130 dart::Precompiler::CompileFunction(dart::Precompiler*, dart::Thread*, dart::Zone*, dart::Function const&)+0x176 pc 0x00000001047f2181 fp 0x000000030d8431d0 dart::Precompiler::ProcessFunction(dart::Function const&)+0xf1 pc 0x00000001047eb60b fp 0x000000030d843230 dart::Precompiler::Iterate()+0xab pc 0x00000001047e59e9 fp 0x000000030d843980 dart::Precompiler::DoCompileAll()+0x1489 pc 0x00000001047e44db fp 0x000000030d843de0 dart::Precompiler::CompileAll()+0xbb pc 0x00000001049d8828 fp 0x000000030d843f50 Dart_Precompile+0x248 pc 0x000000010450ff4b fp 0x000000030d8440e0 dart::bin::main(int, char**)+0x9ab pc 0x0000000205245310 fp 0x000000030d844380 Unknown symbol -- End of DumpStackTrace === Crash occurred when compiling package:photodoc/model/dto/project_dto.dart_ProjectDto_ProjectDto.init in AOT mode in TypePropagation pass *** BEGIN CFG TypePropagation ==== package:photodoc/model/dto/project_dto.dart__$_ProjectDto@1727099614__$_ProjectDto@1727099614. (Constructor) B0[graph]:0 { v0 <- Constant(#-1) T{_Smi} v1 <- Constant(#) T{_OneByteString} v2 <- Constant(#false) T{bool} v3 <- Constant(#null) T{Null?} v4 <- Constant(#_ImmutableList len:0) v5 <- Constant(#_ImmutableList len:0) v6 <- Constant(#_ImmutableList len:0) v7 <- Constant(#_ImmutableList len:0) v8 <- Constant(#true) v5 <- Constant(#<optimized out>) } B3[function entry]:2 { v6 <- Parameter(0) T{_$_ProjectDto} v7 <- SpecialParameter(ArgDescriptor) } StoreField(v6 . id = v0 <int64>) StoreField(v6 . name = v1) StoreField(v6 . eventStoreId = v1) StoreField(v6 . deleted = v2) StoreField(v6 . lastUpdatedByUser = v3) StoreField(v6 . createdDate = v3) StoreField(v6 . createdBy = v3) StoreField(v6 . company = v3) StoreField(v6 . address = v3) StoreField(v6 . customer = v3) StoreField(v6 . distanceToCurrentUser = v1) StoreField(v6 . number = v1) StoreField(v6 . isSynchronizeEnabled = v8) StoreField(v6 . isNewProject = v2) StoreField(v6 . isOffline = v2) StoreField(v6 . _images@1727099614 = v4) StoreField(v6 . _tags@1727099614 = v5) StoreField(v6 . _reports@1727099614 = v6) StoreField(v6 . _participants@1727099614 = v7) Return:470(v3) *** END CFG Dart snapshot generator failed with exit code -6 ===== CRASH ===== si_signo=Segmentation fault: 11(11), si_code=1, si_addr=0xc version=2.19.0 (stable) (Mon Jan 23 11:29:09 2023 -0800) on "macos_x64" pid=59853, thread=-1, isolate_group=isolate(0x7fa106024e00), isolate=(nil)(0x0) os=macos, arch=x64, comp=yes, sim=no isolate_instructions=0, vm_instructions=0 pc 0x0000000102cc627d fp 0x0000000309e04150 dart::FlowGraphTypePropagator::VisitValue(dart::Value*)+0x3d pc 0x0000000102cc5eb7 fp 0x0000000309e041a0 dart::FlowGraphTypePropagator::PropagateRecursive(dart::BlockEntryInstr*)+0xd7 pc 0x0000000102cc5f93 fp 0x0000000309e041f0 dart::FlowGraphTypePropagator::PropagateRecursive(dart::BlockEntryInstr*)+0x1b3 pc 0x0000000102cc5572 fp 0x0000000309e04670 dart::FlowGraphTypePropagator::Propagate()+0x32 pc 0x0000000102cc5508 fp 0x0000000309e04790 dart::FlowGraphTypePropagator::Propagate(dart::FlowGraph*)+0x58 pc 0x0000000102cddcc2 fp 0x0000000309e047b0 dart::CompilerPass_TypePropagation::DoBody(dart::CompilerPassState*) const+0x22 pc 0x0000000102cdc8b0 fp 0x0000000309e04880 dart::CompilerPass::Run(dart::CompilerPassState*) const+0x1b0 pc 0x0000000102cdcada fp 0x0000000309e048b0 dart::CompilerPass::RunInliningPipeline(dart::CompilerPass::PipelineMode, dart::CompilerPassState*)+0x3a pc 0x0000000102c8b533 fp 0x0000000309e04dc0 dart::CallSiteInliner::TryInliningImpl(dart::Function const&, dart::Array const&, dart::InlinedCallData*, bool)+0x1073 pc 0x0000000102c7fa96 fp 0x0000000309e04e60 dart::CallSiteInliner::TryInlining(dart::Function const&, dart::Array const&, dart::InlinedCallData*, bool)+0xf6 pc 0x0000000102c8e88b fp 0x0000000309e04f50 dart::CallSiteInliner::InlineStaticCalls()+0x56b pc 0x0000000102c857ef fp 0x0000000309e05070 dart::CallSiteInliner::InlineCalls()+0x24f pc 0x0000000102c85441 fp 0x0000000309e05170 dart::FlowGraphInliner::Inline()+0x1e1 pc 0x0000000102cddc32 fp 0x0000000309e051e0 dart::CompilerPass_Inlining::DoBody(dart::CompilerPassState*) const+0x52 pc 0x0000000102cdc8b0 fp 0x0000000309e052b0 dart::CompilerPass::Run(dart::CompilerPassState*) const+0x1b0 pc 0x0000000102cdcd99 fp 0x0000000309e052e0 dart::CompilerPass::RunPipeline(dart::CompilerPass::PipelineMode, dart::CompilerPassState*)+0xa9 pc 0x0000000102babbc1 fp 0x0000000309e059b0 dart::PrecompileParsedFunctionHelper::Compile(dart::CompilationPipeline*)+0x611 pc 0x0000000102bac6ba fp 0x0000000309e06030 dart::PrecompileFunctionHelper(dart::Precompiler*, dart::CompilationPipeline*, dart::Function const&, bool)+0x31a pc 0x0000000102ba6a76 fp 0x0000000309e06140 dart::Precompiler::CompileFunction(dart::Precompiler*, dart::Thread*, dart::Zone*, dart::Function const&)+0x176 pc 0x0000000102ba3da1 fp 0x0000000309e061e0 dart::Precompiler::ProcessFunction(dart::Function const&)+0xf1 pc 0x0000000102b9d22b fp 0x0000000309e06240 dart::Precompiler::Iterate()+0xab pc 0x0000000102b97609 fp 0x0000000309e06990 dart::Precompiler::DoCompileAll()+0x1489 pc 0x0000000102b960fb fp 0x0000000309e06df0 dart::Precompiler::CompileAll()+0xbb pc 0x0000000102d96808 fp 0x0000000309e06f60 Dart_Precompile+0x248 pc 0x00000001028d0f4b fp 0x0000000309e070f0 dart::bin::main(int, char**)+0x9ab pc 0x0000000203447310 fp 0x0000000309e07390 Unknown symbol -- End of DumpStackTrace === Crash occurred when compiling package:photodoc/model/dto/project_dto.dart_ProjectDto_ProjectDto.init in AOT mode in TypePropagation pass *** BEGIN CFG TypePropagation ==== package:photodoc/model/dto/project_dto.dart__$_ProjectDto@1727099614__$_ProjectDto@1727099614. (Constructor) B0[graph]:0 { v0 <- Constant(#-1) T{_Smi} v1 <- Constant(#) T{_OneByteString} v2 <- Constant(#false) T{bool} v3 <- Constant(#null) T{Null?} v4 <- Constant(#_ImmutableList len:0) v5 <- Constant(#_ImmutableList len:0) v6 <- Constant(#_ImmutableList len:0) v7 <- Constant(#_ImmutableList len:0) v8 <- Constant(#true) v5 <- Constant(#<optimized out>) } B3[function entry]:2 { v6 <- Parameter(0) T{_$_ProjectDto} v7 <- SpecialParameter(ArgDescriptor) } StoreField(v6 . id = v0 <int64>) StoreField(v6 . name = v1) StoreField(v6 . eventStoreId = v1) StoreField(v6 . deleted = v2) StoreField(v6 . lastUpdatedByUser = v3) StoreField(v6 . createdDate = v3) StoreField(v6 . createdBy = v3) StoreField(v6 . company = v3) StoreField(v6 . address = v3) StoreField(v6 . customer = v3) StoreField(v6 . distanceToCurrentUser = v1) StoreField(v6 . number = v1) StoreField(v6 . isSynchronizeEnabled = v8) StoreField(v6 . isNewProject = v2) StoreField(v6 . isOffline = v2) StoreField(v6 . _images@1727099614 = v4) StoreField(v6 . _tags@1727099614 = v5) StoreField(v6 . _reports@1727099614 = v6) StoreField(v6 . _participants@1727099614 = v7) Return:470(v3) *** END CFG Dart snapshot generator failed with exit code -6 Target android_aot_release_android-arm failed: Exception: AOT snapshotter exited with code -6 Target android_aot_release_android-arm64 failed: Exception: AOT snapshotter exited with code -6 Target android_aot_release_android-x64 failed: Exception: AOT snapshotter exited with code -6 FAILURE: Build failed with an exception. * Where: Script '/Users/sunbro/flutter/packages/flutter_tools/gradle/flutter.gradle' line: 1151 * What went wrong: Execution failed for task ':app:compileFlutterBuildDevRelease'. > Process 'command '/Users/sunbro/flutter/bin/flutter'' finished with non-zero exit value 1 * Try: > Run with --stacktrace option to get the stack trace. > Run with --info or --debug option to get more log output. > Run with --scan to get full insights. * Get more help at https://help.gradle.org BUILD FAILED in 37s ``` <!-- If possible, paste the output of running `flutter doctor -v` here. --> ``` [✓] Flutter (Channel stable, 3.7.0, on macOS 13.1 22C65 darwin-arm64, locale en-DE) • Flutter version 3.7.0 on channel stable at /Users/sunbro/flutter • Upstream repository https://github.com/flutter/flutter.git • Framework revision b06b8b2710 (2 days ago), 2023-01-23 16:55:55 -0800 • Engine revision b24591ed32 • Dart version 2.19.0 • DevTools version 2.20.1 [✓] Android toolchain - develop for Android devices (Android SDK version 33.0.1) • Android SDK at /Users/sunbro/Library/Android/sdk • Platform android-33, build-tools 33.0.1 • Java binary at: /Applications/Android Studio.app/Contents/jre/Contents/Home/bin/java • Java version OpenJDK Runtime Environment (build 11.0.15+0-b2043.56-8887301) • All Android licenses accepted. [✓] Xcode - develop for iOS and macOS (Xcode 14.2) • Xcode at /Applications/Xcode.app/Contents/Developer • Build 14C18 • CocoaPods version 1.11.3 [✓] Chrome - develop for the web • Chrome at /Applications/Google Chrome.app/Contents/MacOS/Google Chrome [✓] Android Studio (version 2022.1) • Android Studio at /Applications/Android Studio.app/Contents • Flutter plugin can be installed from: 🔨 https://plugins.jetbrains.com/plugin/9212-flutter • Dart plugin can be installed from: 🔨 https://plugins.jetbrains.com/plugin/6351-dart • Java version OpenJDK Runtime Environment (build 11.0.15+0-b2043.56-8887301) [✓] IntelliJ IDEA Community Edition (version 2022.2) • IntelliJ at /Applications/IntelliJ IDEA CE.app • Flutter plugin version 70.0.5 • Dart plugin version 222.3345.108 [✓] Connected device (2 available) • macOS (desktop) • macos • darwin-arm64 • macOS 13.1 22C65 darwin-arm64 • Chrome (web) • chrome • web-javascript • Google Chrome 109.0.5414.119 [✓] HTTP Host Availability • All required HTTP hosts are available • No issues found! ``` </details>
1.0
flutter build --release crashes 3.7.0 - <!-- Thank you for using Flutter! If you are looking for support, please check out our documentation or consider asking a question on Stack Overflow: * https://flutter.dev/ * https://api.flutter.dev/ * https://stackoverflow.com/questions/tagged/flutter?sort=frequent If you have found a bug or if our documentation doesn't have an answer to what you're looking for, then fill out the template below. Please read our guide to filing a bug first: https://flutter.dev/docs/resources/bug-reports --> ## Steps to Reproduce <!-- Please tell us exactly how to reproduce the problem you are running into. --> Building a debug app works `flutter build apk --debug --flavor dev -t lib/main.dart` ```sh Running Gradle task 'assembleDevDebug'... 33.5s ✓ Built build/app/outputs/flutter-apk/app-dev-debug.apk. ``` But building for release crashes `flutter build apk --release --flavor dev -t lib/main.dart` ```sh ===== CRASH ===== si_signo=Segmentation fault: 11(11), si_code=1, si_addr=0xc version=2.19.0 (stable) (Mon Jan 23 11:29:09 2023 -0800) on "macos_x64" pid=61474, thread=-1, isolate_group=isolate(0x7fec27808200), isolate=(nil)(0x0) os=macos, arch=x64, comp=yes, sim=no isolate_instructions=0, vm_instructions=0 ``` Before with Flutter 3.3.10 --release build successful. This is a App in released in production for Android and iOS ## Logs <details> <summary>Logs</summary> <!-- Include the full logs of the commands you are running between the lines with the backticks below. If you are running any "flutter" commands, please include the output of running them with "--verbose"; for example, the output of running "flutter --verbose create foo". --> ``` 💪 Building with sound null safety 💪 Expected to find fonts for (packages/cupertino_icons/CupertinoIcons, MaterialIcons, WuerthIcon), but found (MaterialIcons, WuerthIcon). This usually means you are referring to font families in an IconData class but not including them in the assets section of your pubspec.yaml, are missing the package that would include them, or are missing "uses-material-design: true". ===== CRASH ===== si_signo=Segmentation fault: 11(11), si_code=1, si_addr=0xc version=2.19.0 (stable) (Mon Jan 23 11:29:09 2023 -0800) on "macos_arm" pid=59850, thread=-1, isolate_group=isolate(0x7fe1b801b400), isolate=(nil)(0x0) os=macos, arch=arm, comp=no, sim=no isolate_instructions=0, vm_instructions=0 pc 0x000000010139e2ad fp 0x0000000306bfe100 dart::FlowGraphTypePropagator::VisitValue(dart::Value*)+0x3d pc 0x000000010139dee7 fp 0x0000000306bfe150 dart::FlowGraphTypePropagator::PropagateRecursive(dart::BlockEntryInstr*)+0xd7 pc 0x000000010139dfc3 fp 0x0000000306bfe1a0 dart::FlowGraphTypePropagator::PropagateRecursive(dart::BlockEntryInstr*)+0x1b3 pc 0x000000010139d5a2 fp 0x0000000306bfe620 dart::FlowGraphTypePropagator::Propagate()+0x32 pc 0x000000010139d538 fp 0x0000000306bfe740 dart::FlowGraphTypePropagator::Propagate(dart::FlowGraph*)+0x58 pc 0x00000001013b5662 fp 0x0000000306bfe760 dart::CompilerPass_TypePropagation::DoBody(dart::CompilerPassState*) const+0x22 pc 0x00000001013b4250 fp 0x0000000306bfe830 dart::CompilerPass::Run(dart::CompilerPassState*) const+0x1b0 pc 0x00000001013b447a fp 0x0000000306bfe860 dart::CompilerPass::RunInliningPipeline(dart::CompilerPass::PipelineMode, dart::CompilerPassState*)+0x3a pc 0x00000001013630fb fp 0x0000000306bfed70 dart::CallSiteInliner::TryInliningImpl(dart::Function const&, dart::Array const&, dart::InlinedCallData*, bool)+0x100b pc 0x00000001013577d6 fp 0x0000000306bfee10 dart::CallSiteInliner::TryInlining(dart::Function const&, dart::Array const&, dart::InlinedCallData*, bool)+0xf6 pc 0x00000001013662db fp 0x0000000306bfef00 dart::CallSiteInliner::InlineStaticCalls()+0x56b pc 0x000000010135d3df fp 0x0000000306bff020 dart::CallSiteInliner::InlineCalls()+0x24f pc 0x000000010135d031 fp 0x0000000306bff120 dart::FlowGraphInliner::Inline()+0x1e1 pc 0x00000001013b55d2 fp 0x0000000306bff190 dart::CompilerPass_Inlining::DoBody(dart::CompilerPassState*) const+0x52 pc 0x00000001013b4250 fp 0x0000000306bff260 dart::CompilerPass::Run(dart::CompilerPassState*) const+0x1b0 pc 0x00000001013b4739 fp 0x0000000306bff290 dart::CompilerPass::RunPipeline(dart::CompilerPass::PipelineMode, dart::CompilerPassState*)+0xa9 pc 0x0000000101286bf9 fp 0x0000000306bff960 dart::PrecompileParsedFunctionHelper::Compile(dart::CompilationPipeline*)+0x5c9 pc 0x00000001012876fd fp 0x0000000306bfffe0 dart::PrecompileFunctionHelper(dart::Precompiler*, dart::CompilationPipeline*, dart::Function const&, bool)+0x31d pc 0x00000001012824f6 fp 0x0000000306c000f0 dart::Precompiler::CompileFunction(dart::Precompiler*, dart::Thread*, dart::Zone*, dart::Function const&)+0x176 pc 0x000000010127fd81 fp 0x0000000306c00190 dart::Precompiler::ProcessFunction(dart::Function const&)+0xf1 pc 0x000000010127a054 fp 0x0000000306c001f0 dart::Precompiler::Iterate()+0x94 pc 0x0000000101274b19 fp 0x0000000306c00940 dart::Precompiler::DoCompileAll()+0x1489 pc 0x000000010127360b fp 0x0000000306c00da0 dart::Precompiler::CompileAll()+0xbb pc 0x0000000101468ab8 fp 0x0000000306c00f10 Dart_Precompile+0x248 pc 0x0000000100fc2f4b fp 0x0000000306c010a0 dart::bin::main(int, char**)+0x9ab pc 0x0000000201b4f310 fp 0x0000000306c01340 Unknown symbol -- End of DumpStackTrace === Crash occurred when compiling package:photodoc/model/dto/project_dto.dart_ProjectDto_ProjectDto.init in AOT mode in TypePropagation pass *** BEGIN CFG TypePropagation ==== package:photodoc/model/dto/project_dto.dart__$_ProjectDto@1727099614__$_ProjectDto@1727099614. (Constructor) B0[graph]:0 { v0 <- Constant(#-1) T{_Smi} v1 <- Constant(#) T{_OneByteString} v2 <- Constant(#false) T{bool} v3 <- Constant(#null) T{Null?} v4 <- Constant(#_ImmutableList len:0) v5 <- Constant(#_ImmutableList len:0) v6 <- Constant(#_ImmutableList len:0) v7 <- Constant(#_ImmutableList len:0) v8 <- Constant(#true) v5 <- Constant(#<optimized out>) } B3[function entry]:2 { v6 <- Parameter(0) T{_$_ProjectDto} v7 <- SpecialParameter(ArgDescriptor) } StoreField(v6 . id = v0 <int64>) StoreField(v6 . name = v1) StoreField(v6 . eventStoreId = v1) StoreField(v6 . deleted = v2) StoreField(v6 . lastUpdatedByUser = v3) StoreField(v6 . createdDate = v3) StoreField(v6 . createdBy = v3) StoreField(v6 . company = v3) StoreField(v6 . address = v3) StoreField(v6 . customer = v3) StoreField(v6 . distanceToCurrentUser = v1) StoreField(v6 . number = v1) StoreField(v6 . isSynchronizeEnabled = v8) StoreField(v6 . isNewProject = v2) StoreField(v6 . isOffline = v2) StoreField(v6 . _images@1727099614 = v4) StoreField(v6 . _tags@1727099614 = v5) StoreField(v6 . _reports@1727099614 = v6) StoreField(v6 . _participants@1727099614 = v7) Return:470(v3) *** END CFG Dart snapshot generator failed with exit code -6 ===== CRASH ===== si_signo=Segmentation fault: 11(11), si_code=1, si_addr=0xc version=2.19.0 (stable) (Mon Jan 23 11:29:09 2023 -0800) on "macos_simarm64" pid=59851, thread=-1, isolate_group=isolate(0x7fd545008200), isolate=(nil)(0x0) os=macos, arch=arm64, comp=yes, sim=yes isolate_instructions=0, vm_instructions=0 pc 0x000000010490b1cd fp 0x000000030d841140 dart::FlowGraphTypePropagator::VisitValue(dart::Value*)+0x3d pc 0x000000010490ae07 fp 0x000000030d841190 dart::FlowGraphTypePropagator::PropagateRecursive(dart::BlockEntryInstr*)+0xd7 pc 0x000000010490aee3 fp 0x000000030d8411e0 dart::FlowGraphTypePropagator::PropagateRecursive(dart::BlockEntryInstr*)+0x1b3 pc 0x000000010490a4c2 fp 0x000000030d841660 dart::FlowGraphTypePropagator::Propagate()+0x32 pc 0x000000010490a458 fp 0x000000030d841780 dart::FlowGraphTypePropagator::Propagate(dart::FlowGraph*)+0x58 pc 0x0000000104922c12 fp 0x000000030d8417a0 dart::CompilerPass_TypePropagation::DoBody(dart::CompilerPassState*) const+0x22 pc 0x0000000104921800 fp 0x000000030d841870 dart::CompilerPass::Run(dart::CompilerPassState*) const+0x1b0 pc 0x0000000104921a2a fp 0x000000030d8418a0 dart::CompilerPass::RunInliningPipeline(dart::CompilerPass::PipelineMode, dart::CompilerPassState*)+0x3a pc 0x00000001048d0393 fp 0x000000030d841db0 dart::CallSiteInliner::TryInliningImpl(dart::Function const&, dart::Array const&, dart::InlinedCallData*, bool)+0x1073 pc 0x00000001048c48f6 fp 0x000000030d841e50 dart::CallSiteInliner::TryInlining(dart::Function const&, dart::Array const&, dart::InlinedCallData*, bool)+0xf6 pc 0x00000001048d36eb fp 0x000000030d841f40 dart::CallSiteInliner::InlineStaticCalls()+0x56b pc 0x00000001048ca64f fp 0x000000030d842060 dart::CallSiteInliner::InlineCalls()+0x24f pc 0x00000001048ca2a1 fp 0x000000030d842160 dart::FlowGraphInliner::Inline()+0x1e1 pc 0x0000000104922b82 fp 0x000000030d8421d0 dart::CompilerPass_Inlining::DoBody(dart::CompilerPassState*) const+0x52 pc 0x0000000104921800 fp 0x000000030d8422a0 dart::CompilerPass::Run(dart::CompilerPassState*) const+0x1b0 pc 0x0000000104921ce9 fp 0x000000030d8422d0 dart::CompilerPass::RunPipeline(dart::CompilerPass::PipelineMode, dart::CompilerPassState*)+0xa9 pc 0x00000001047f9fa1 fp 0x000000030d8429a0 dart::PrecompileParsedFunctionHelper::Compile(dart::CompilationPipeline*)+0x611 pc 0x00000001047faa9a fp 0x000000030d843020 dart::PrecompileFunctionHelper(dart::Precompiler*, dart::CompilationPipeline*, dart::Function const&, bool)+0x31a pc 0x00000001047f4e56 fp 0x000000030d843130 dart::Precompiler::CompileFunction(dart::Precompiler*, dart::Thread*, dart::Zone*, dart::Function const&)+0x176 pc 0x00000001047f2181 fp 0x000000030d8431d0 dart::Precompiler::ProcessFunction(dart::Function const&)+0xf1 pc 0x00000001047eb60b fp 0x000000030d843230 dart::Precompiler::Iterate()+0xab pc 0x00000001047e59e9 fp 0x000000030d843980 dart::Precompiler::DoCompileAll()+0x1489 pc 0x00000001047e44db fp 0x000000030d843de0 dart::Precompiler::CompileAll()+0xbb pc 0x00000001049d8828 fp 0x000000030d843f50 Dart_Precompile+0x248 pc 0x000000010450ff4b fp 0x000000030d8440e0 dart::bin::main(int, char**)+0x9ab pc 0x0000000205245310 fp 0x000000030d844380 Unknown symbol -- End of DumpStackTrace === Crash occurred when compiling package:photodoc/model/dto/project_dto.dart_ProjectDto_ProjectDto.init in AOT mode in TypePropagation pass *** BEGIN CFG TypePropagation ==== package:photodoc/model/dto/project_dto.dart__$_ProjectDto@1727099614__$_ProjectDto@1727099614. (Constructor) B0[graph]:0 { v0 <- Constant(#-1) T{_Smi} v1 <- Constant(#) T{_OneByteString} v2 <- Constant(#false) T{bool} v3 <- Constant(#null) T{Null?} v4 <- Constant(#_ImmutableList len:0) v5 <- Constant(#_ImmutableList len:0) v6 <- Constant(#_ImmutableList len:0) v7 <- Constant(#_ImmutableList len:0) v8 <- Constant(#true) v5 <- Constant(#<optimized out>) } B3[function entry]:2 { v6 <- Parameter(0) T{_$_ProjectDto} v7 <- SpecialParameter(ArgDescriptor) } StoreField(v6 . id = v0 <int64>) StoreField(v6 . name = v1) StoreField(v6 . eventStoreId = v1) StoreField(v6 . deleted = v2) StoreField(v6 . lastUpdatedByUser = v3) StoreField(v6 . createdDate = v3) StoreField(v6 . createdBy = v3) StoreField(v6 . company = v3) StoreField(v6 . address = v3) StoreField(v6 . customer = v3) StoreField(v6 . distanceToCurrentUser = v1) StoreField(v6 . number = v1) StoreField(v6 . isSynchronizeEnabled = v8) StoreField(v6 . isNewProject = v2) StoreField(v6 . isOffline = v2) StoreField(v6 . _images@1727099614 = v4) StoreField(v6 . _tags@1727099614 = v5) StoreField(v6 . _reports@1727099614 = v6) StoreField(v6 . _participants@1727099614 = v7) Return:470(v3) *** END CFG Dart snapshot generator failed with exit code -6 ===== CRASH ===== si_signo=Segmentation fault: 11(11), si_code=1, si_addr=0xc version=2.19.0 (stable) (Mon Jan 23 11:29:09 2023 -0800) on "macos_x64" pid=59853, thread=-1, isolate_group=isolate(0x7fa106024e00), isolate=(nil)(0x0) os=macos, arch=x64, comp=yes, sim=no isolate_instructions=0, vm_instructions=0 pc 0x0000000102cc627d fp 0x0000000309e04150 dart::FlowGraphTypePropagator::VisitValue(dart::Value*)+0x3d pc 0x0000000102cc5eb7 fp 0x0000000309e041a0 dart::FlowGraphTypePropagator::PropagateRecursive(dart::BlockEntryInstr*)+0xd7 pc 0x0000000102cc5f93 fp 0x0000000309e041f0 dart::FlowGraphTypePropagator::PropagateRecursive(dart::BlockEntryInstr*)+0x1b3 pc 0x0000000102cc5572 fp 0x0000000309e04670 dart::FlowGraphTypePropagator::Propagate()+0x32 pc 0x0000000102cc5508 fp 0x0000000309e04790 dart::FlowGraphTypePropagator::Propagate(dart::FlowGraph*)+0x58 pc 0x0000000102cddcc2 fp 0x0000000309e047b0 dart::CompilerPass_TypePropagation::DoBody(dart::CompilerPassState*) const+0x22 pc 0x0000000102cdc8b0 fp 0x0000000309e04880 dart::CompilerPass::Run(dart::CompilerPassState*) const+0x1b0 pc 0x0000000102cdcada fp 0x0000000309e048b0 dart::CompilerPass::RunInliningPipeline(dart::CompilerPass::PipelineMode, dart::CompilerPassState*)+0x3a pc 0x0000000102c8b533 fp 0x0000000309e04dc0 dart::CallSiteInliner::TryInliningImpl(dart::Function const&, dart::Array const&, dart::InlinedCallData*, bool)+0x1073 pc 0x0000000102c7fa96 fp 0x0000000309e04e60 dart::CallSiteInliner::TryInlining(dart::Function const&, dart::Array const&, dart::InlinedCallData*, bool)+0xf6 pc 0x0000000102c8e88b fp 0x0000000309e04f50 dart::CallSiteInliner::InlineStaticCalls()+0x56b pc 0x0000000102c857ef fp 0x0000000309e05070 dart::CallSiteInliner::InlineCalls()+0x24f pc 0x0000000102c85441 fp 0x0000000309e05170 dart::FlowGraphInliner::Inline()+0x1e1 pc 0x0000000102cddc32 fp 0x0000000309e051e0 dart::CompilerPass_Inlining::DoBody(dart::CompilerPassState*) const+0x52 pc 0x0000000102cdc8b0 fp 0x0000000309e052b0 dart::CompilerPass::Run(dart::CompilerPassState*) const+0x1b0 pc 0x0000000102cdcd99 fp 0x0000000309e052e0 dart::CompilerPass::RunPipeline(dart::CompilerPass::PipelineMode, dart::CompilerPassState*)+0xa9 pc 0x0000000102babbc1 fp 0x0000000309e059b0 dart::PrecompileParsedFunctionHelper::Compile(dart::CompilationPipeline*)+0x611 pc 0x0000000102bac6ba fp 0x0000000309e06030 dart::PrecompileFunctionHelper(dart::Precompiler*, dart::CompilationPipeline*, dart::Function const&, bool)+0x31a pc 0x0000000102ba6a76 fp 0x0000000309e06140 dart::Precompiler::CompileFunction(dart::Precompiler*, dart::Thread*, dart::Zone*, dart::Function const&)+0x176 pc 0x0000000102ba3da1 fp 0x0000000309e061e0 dart::Precompiler::ProcessFunction(dart::Function const&)+0xf1 pc 0x0000000102b9d22b fp 0x0000000309e06240 dart::Precompiler::Iterate()+0xab pc 0x0000000102b97609 fp 0x0000000309e06990 dart::Precompiler::DoCompileAll()+0x1489 pc 0x0000000102b960fb fp 0x0000000309e06df0 dart::Precompiler::CompileAll()+0xbb pc 0x0000000102d96808 fp 0x0000000309e06f60 Dart_Precompile+0x248 pc 0x00000001028d0f4b fp 0x0000000309e070f0 dart::bin::main(int, char**)+0x9ab pc 0x0000000203447310 fp 0x0000000309e07390 Unknown symbol -- End of DumpStackTrace === Crash occurred when compiling package:photodoc/model/dto/project_dto.dart_ProjectDto_ProjectDto.init in AOT mode in TypePropagation pass *** BEGIN CFG TypePropagation ==== package:photodoc/model/dto/project_dto.dart__$_ProjectDto@1727099614__$_ProjectDto@1727099614. (Constructor) B0[graph]:0 { v0 <- Constant(#-1) T{_Smi} v1 <- Constant(#) T{_OneByteString} v2 <- Constant(#false) T{bool} v3 <- Constant(#null) T{Null?} v4 <- Constant(#_ImmutableList len:0) v5 <- Constant(#_ImmutableList len:0) v6 <- Constant(#_ImmutableList len:0) v7 <- Constant(#_ImmutableList len:0) v8 <- Constant(#true) v5 <- Constant(#<optimized out>) } B3[function entry]:2 { v6 <- Parameter(0) T{_$_ProjectDto} v7 <- SpecialParameter(ArgDescriptor) } StoreField(v6 . id = v0 <int64>) StoreField(v6 . name = v1) StoreField(v6 . eventStoreId = v1) StoreField(v6 . deleted = v2) StoreField(v6 . lastUpdatedByUser = v3) StoreField(v6 . createdDate = v3) StoreField(v6 . createdBy = v3) StoreField(v6 . company = v3) StoreField(v6 . address = v3) StoreField(v6 . customer = v3) StoreField(v6 . distanceToCurrentUser = v1) StoreField(v6 . number = v1) StoreField(v6 . isSynchronizeEnabled = v8) StoreField(v6 . isNewProject = v2) StoreField(v6 . isOffline = v2) StoreField(v6 . _images@1727099614 = v4) StoreField(v6 . _tags@1727099614 = v5) StoreField(v6 . _reports@1727099614 = v6) StoreField(v6 . _participants@1727099614 = v7) Return:470(v3) *** END CFG Dart snapshot generator failed with exit code -6 Target android_aot_release_android-arm failed: Exception: AOT snapshotter exited with code -6 Target android_aot_release_android-arm64 failed: Exception: AOT snapshotter exited with code -6 Target android_aot_release_android-x64 failed: Exception: AOT snapshotter exited with code -6 FAILURE: Build failed with an exception. * Where: Script '/Users/sunbro/flutter/packages/flutter_tools/gradle/flutter.gradle' line: 1151 * What went wrong: Execution failed for task ':app:compileFlutterBuildDevRelease'. > Process 'command '/Users/sunbro/flutter/bin/flutter'' finished with non-zero exit value 1 * Try: > Run with --stacktrace option to get the stack trace. > Run with --info or --debug option to get more log output. > Run with --scan to get full insights. * Get more help at https://help.gradle.org BUILD FAILED in 37s ``` <!-- If possible, paste the output of running `flutter doctor -v` here. --> ``` [✓] Flutter (Channel stable, 3.7.0, on macOS 13.1 22C65 darwin-arm64, locale en-DE) • Flutter version 3.7.0 on channel stable at /Users/sunbro/flutter • Upstream repository https://github.com/flutter/flutter.git • Framework revision b06b8b2710 (2 days ago), 2023-01-23 16:55:55 -0800 • Engine revision b24591ed32 • Dart version 2.19.0 • DevTools version 2.20.1 [✓] Android toolchain - develop for Android devices (Android SDK version 33.0.1) • Android SDK at /Users/sunbro/Library/Android/sdk • Platform android-33, build-tools 33.0.1 • Java binary at: /Applications/Android Studio.app/Contents/jre/Contents/Home/bin/java • Java version OpenJDK Runtime Environment (build 11.0.15+0-b2043.56-8887301) • All Android licenses accepted. [✓] Xcode - develop for iOS and macOS (Xcode 14.2) • Xcode at /Applications/Xcode.app/Contents/Developer • Build 14C18 • CocoaPods version 1.11.3 [✓] Chrome - develop for the web • Chrome at /Applications/Google Chrome.app/Contents/MacOS/Google Chrome [✓] Android Studio (version 2022.1) • Android Studio at /Applications/Android Studio.app/Contents • Flutter plugin can be installed from: 🔨 https://plugins.jetbrains.com/plugin/9212-flutter • Dart plugin can be installed from: 🔨 https://plugins.jetbrains.com/plugin/6351-dart • Java version OpenJDK Runtime Environment (build 11.0.15+0-b2043.56-8887301) [✓] IntelliJ IDEA Community Edition (version 2022.2) • IntelliJ at /Applications/IntelliJ IDEA CE.app • Flutter plugin version 70.0.5 • Dart plugin version 222.3345.108 [✓] Connected device (2 available) • macOS (desktop) • macos • darwin-arm64 • macOS 13.1 22C65 darwin-arm64 • Chrome (web) • chrome • web-javascript • Google Chrome 109.0.5414.119 [✓] HTTP Host Availability • All required HTTP hosts are available • No issues found! ``` </details>
non_process
flutter build release crashes thank you for using flutter if you are looking for support please check out our documentation or consider asking a question on stack overflow if you have found a bug or if our documentation doesn t have an answer to what you re looking for then fill out the template below please read our guide to filing a bug first steps to reproduce building a debug app works flutter build apk debug flavor dev t lib main dart sh running gradle task assembledevdebug ✓ built build app outputs flutter apk app dev debug apk but building for release crashes flutter build apk release flavor dev t lib main dart sh crash si signo segmentation fault si code si addr version stable mon jan on macos pid thread isolate group isolate isolate nil os macos arch comp yes sim no isolate instructions vm instructions before with flutter release build successful this is a app in released in production for android and ios logs logs include the full logs of the commands you are running between the lines with the backticks below if you are running any flutter commands please include the output of running them with verbose for example the output of running flutter verbose create foo 💪 building with sound null safety 💪 expected to find fonts for packages cupertino icons cupertinoicons materialicons wuerthicon but found materialicons wuerthicon this usually means you are referring to font families in an icondata class but not including them in the assets section of your pubspec yaml are missing the package that would include them or are missing uses material design true crash si signo segmentation fault si code si addr version stable mon jan on macos arm pid thread isolate group isolate isolate nil os macos arch arm comp no sim no isolate instructions vm instructions pc fp dart flowgraphtypepropagator visitvalue dart value pc fp dart flowgraphtypepropagator propagaterecursive dart blockentryinstr pc fp dart flowgraphtypepropagator propagaterecursive dart blockentryinstr pc fp dart flowgraphtypepropagator propagate pc fp dart flowgraphtypepropagator propagate dart flowgraph pc fp dart compilerpass typepropagation dobody dart compilerpassstate const pc fp dart compilerpass run dart compilerpassstate const pc fp dart compilerpass runinliningpipeline dart compilerpass pipelinemode dart compilerpassstate pc fp dart callsiteinliner tryinliningimpl dart function const dart array const dart inlinedcalldata bool pc fp dart callsiteinliner tryinlining dart function const dart array const dart inlinedcalldata bool pc fp dart callsiteinliner inlinestaticcalls pc fp dart callsiteinliner inlinecalls pc fp dart flowgraphinliner inline pc fp dart compilerpass inlining dobody dart compilerpassstate const pc fp dart compilerpass run dart compilerpassstate const pc fp dart compilerpass runpipeline dart compilerpass pipelinemode dart compilerpassstate pc fp dart precompileparsedfunctionhelper compile dart compilationpipeline pc fp dart precompilefunctionhelper dart precompiler dart compilationpipeline dart function const bool pc fp dart precompiler compilefunction dart precompiler dart thread dart zone dart function const pc fp dart precompiler processfunction dart function const pc fp dart precompiler iterate pc fp dart precompiler docompileall pc fp dart precompiler compileall pc fp dart precompile pc fp dart bin main int char pc fp unknown symbol end of dumpstacktrace crash occurred when compiling package photodoc model dto project dto dart projectdto projectdto init in aot mode in typepropagation pass begin cfg typepropagation package photodoc model dto project dto dart projectdto projectdto constructor constant t smi constant t onebytestring constant false t bool constant null t null constant immutablelist len constant immutablelist len constant immutablelist len constant immutablelist len constant true parameter t projectdto specialparameter argdescriptor storefield id storefield name storefield eventstoreid storefield deleted storefield lastupdatedbyuser storefield createddate storefield createdby storefield company storefield address storefield customer storefield distancetocurrentuser storefield number storefield issynchronizeenabled storefield isnewproject storefield isoffline storefield images storefield tags storefield reports storefield participants return end cfg dart snapshot generator failed with exit code crash si signo segmentation fault si code si addr version stable mon jan on macos pid thread isolate group isolate isolate nil os macos arch comp yes sim yes isolate instructions vm instructions pc fp dart flowgraphtypepropagator visitvalue dart value pc fp dart flowgraphtypepropagator propagaterecursive dart blockentryinstr pc fp dart flowgraphtypepropagator propagaterecursive dart blockentryinstr pc fp dart flowgraphtypepropagator propagate pc fp dart flowgraphtypepropagator propagate dart flowgraph pc fp dart compilerpass typepropagation dobody dart compilerpassstate const pc fp dart compilerpass run dart compilerpassstate const pc fp dart compilerpass runinliningpipeline dart compilerpass pipelinemode dart compilerpassstate pc fp dart callsiteinliner tryinliningimpl dart function const dart array const dart inlinedcalldata bool pc fp dart callsiteinliner tryinlining dart function const dart array const dart inlinedcalldata bool pc fp dart callsiteinliner inlinestaticcalls pc fp dart callsiteinliner inlinecalls pc fp dart flowgraphinliner inline pc fp dart compilerpass inlining dobody dart compilerpassstate const pc fp dart compilerpass run dart compilerpassstate const pc fp dart compilerpass runpipeline dart compilerpass pipelinemode dart compilerpassstate pc fp dart precompileparsedfunctionhelper compile dart compilationpipeline pc fp dart precompilefunctionhelper dart precompiler dart compilationpipeline dart function const bool pc fp dart precompiler compilefunction dart precompiler dart thread dart zone dart function const pc fp dart precompiler processfunction dart function const pc fp dart precompiler iterate pc fp dart precompiler docompileall pc fp dart precompiler compileall pc fp dart precompile pc fp dart bin main int char pc fp unknown symbol end of dumpstacktrace crash occurred when compiling package photodoc model dto project dto dart projectdto projectdto init in aot mode in typepropagation pass begin cfg typepropagation package photodoc model dto project dto dart projectdto projectdto constructor constant t smi constant t onebytestring constant false t bool constant null t null constant immutablelist len constant immutablelist len constant immutablelist len constant immutablelist len constant true parameter t projectdto specialparameter argdescriptor storefield id storefield name storefield eventstoreid storefield deleted storefield lastupdatedbyuser storefield createddate storefield createdby storefield company storefield address storefield customer storefield distancetocurrentuser storefield number storefield issynchronizeenabled storefield isnewproject storefield isoffline storefield images storefield tags storefield reports storefield participants return end cfg dart snapshot generator failed with exit code crash si signo segmentation fault si code si addr version stable mon jan on macos pid thread isolate group isolate isolate nil os macos arch comp yes sim no isolate instructions vm instructions pc fp dart flowgraphtypepropagator visitvalue dart value pc fp dart flowgraphtypepropagator propagaterecursive dart blockentryinstr pc fp dart flowgraphtypepropagator propagaterecursive dart blockentryinstr pc fp dart flowgraphtypepropagator propagate pc fp dart flowgraphtypepropagator propagate dart flowgraph pc fp dart compilerpass typepropagation dobody dart compilerpassstate const pc fp dart compilerpass run dart compilerpassstate const pc fp dart compilerpass runinliningpipeline dart compilerpass pipelinemode dart compilerpassstate pc fp dart callsiteinliner tryinliningimpl dart function const dart array const dart inlinedcalldata bool pc fp dart callsiteinliner tryinlining dart function const dart array const dart inlinedcalldata bool pc fp dart callsiteinliner inlinestaticcalls pc fp dart callsiteinliner inlinecalls pc fp dart flowgraphinliner inline pc fp dart compilerpass inlining dobody dart compilerpassstate const pc fp dart compilerpass run dart compilerpassstate const pc fp dart compilerpass runpipeline dart compilerpass pipelinemode dart compilerpassstate pc fp dart precompileparsedfunctionhelper compile dart compilationpipeline pc fp dart precompilefunctionhelper dart precompiler dart compilationpipeline dart function const bool pc fp dart precompiler compilefunction dart precompiler dart thread dart zone dart function const pc fp dart precompiler processfunction dart function const pc fp dart precompiler iterate pc fp dart precompiler docompileall pc fp dart precompiler compileall pc fp dart precompile pc fp dart bin main int char pc fp unknown symbol end of dumpstacktrace crash occurred when compiling package photodoc model dto project dto dart projectdto projectdto init in aot mode in typepropagation pass begin cfg typepropagation package photodoc model dto project dto dart projectdto projectdto constructor constant t smi constant t onebytestring constant false t bool constant null t null constant immutablelist len constant immutablelist len constant immutablelist len constant immutablelist len constant true parameter t projectdto specialparameter argdescriptor storefield id storefield name storefield eventstoreid storefield deleted storefield lastupdatedbyuser storefield createddate storefield createdby storefield company storefield address storefield customer storefield distancetocurrentuser storefield number storefield issynchronizeenabled storefield isnewproject storefield isoffline storefield images storefield tags storefield reports storefield participants return end cfg dart snapshot generator failed with exit code target android aot release android arm failed exception aot snapshotter exited with code target android aot release android failed exception aot snapshotter exited with code target android aot release android failed exception aot snapshotter exited with code failure build failed with an exception where script users sunbro flutter packages flutter tools gradle flutter gradle line what went wrong execution failed for task app compileflutterbuilddevrelease process command users sunbro flutter bin flutter finished with non zero exit value try run with stacktrace option to get the stack trace run with info or debug option to get more log output run with scan to get full insights get more help at build failed in flutter channel stable on macos darwin locale en de • flutter version on channel stable at users sunbro flutter • upstream repository • framework revision days ago • engine revision • dart version • devtools version android toolchain develop for android devices android sdk version • android sdk at users sunbro library android sdk • platform android build tools • java binary at applications android studio app contents jre contents home bin java • java version openjdk runtime environment build • all android licenses accepted xcode develop for ios and macos xcode • xcode at applications xcode app contents developer • build • cocoapods version chrome develop for the web • chrome at applications google chrome app contents macos google chrome android studio version • android studio at applications android studio app contents • flutter plugin can be installed from 🔨 • dart plugin can be installed from 🔨 • java version openjdk runtime environment build intellij idea community edition version • intellij at applications intellij idea ce app • flutter plugin version • dart plugin version connected device available • macos desktop • macos • darwin • macos darwin • chrome web • chrome • web javascript • google chrome http host availability • all required http hosts are available • no issues found
0
35,611
2,791,544,550
IssuesEvent
2015-05-10 07:23:21
tgstation/-tg-station
https://api.github.com/repos/tgstation/-tg-station
closed
Occasional massive power drain in all grids.
Blame MSO Bug Priority: CRITICAL
I'm not sure how to reproduce this but a round happened earlier where everywhere started losing power. SMES were outputting, APCs said they had a grid connection but wouldn't charge. During this round, the admins were spammed with a lot of emitters lost power messages, more-so than usual. The power usage on SMES outputs were showing red as if it was hitting their set maximum. Mining and the AI sat APCs ran out of power and shut off. By the time it was reported the round ended before I had a change to investigate.
1.0
Occasional massive power drain in all grids. - I'm not sure how to reproduce this but a round happened earlier where everywhere started losing power. SMES were outputting, APCs said they had a grid connection but wouldn't charge. During this round, the admins were spammed with a lot of emitters lost power messages, more-so than usual. The power usage on SMES outputs were showing red as if it was hitting their set maximum. Mining and the AI sat APCs ran out of power and shut off. By the time it was reported the round ended before I had a change to investigate.
non_process
occasional massive power drain in all grids i m not sure how to reproduce this but a round happened earlier where everywhere started losing power smes were outputting apcs said they had a grid connection but wouldn t charge during this round the admins were spammed with a lot of emitters lost power messages more so than usual the power usage on smes outputs were showing red as if it was hitting their set maximum mining and the ai sat apcs ran out of power and shut off by the time it was reported the round ended before i had a change to investigate
0
11,719
14,547,918,253
IssuesEvent
2020-12-16 00:01:18
pacificclimate/quail
https://api.github.com/repos/pacificclimate/quail
closed
Maximum Consecutive Wet Days
process
## Description This function takes a climdexInput object as input and computes the climdex index CWD: the annual maximum length of wet spells, in days. Wet spells are considered to be sequences of days where daily precipitation is at least 1mm per day. ## Function to wrap [`climdex.cwd`](https://github.com/pacificclimate/climdex.pcic/blob/master/R/climdex.r#L1223)
1.0
Maximum Consecutive Wet Days - ## Description This function takes a climdexInput object as input and computes the climdex index CWD: the annual maximum length of wet spells, in days. Wet spells are considered to be sequences of days where daily precipitation is at least 1mm per day. ## Function to wrap [`climdex.cwd`](https://github.com/pacificclimate/climdex.pcic/blob/master/R/climdex.r#L1223)
process
maximum consecutive wet days description this function takes a climdexinput object as input and computes the climdex index cwd the annual maximum length of wet spells in days wet spells are considered to be sequences of days where daily precipitation is at least per day function to wrap
1
55,170
13,979,788,254
IssuesEvent
2020-10-27 01:04:05
jgeraigery/LocalCatalogManager
https://api.github.com/repos/jgeraigery/LocalCatalogManager
opened
CVE-2020-7020 (Low) detected in elasticsearch-1.4.4.jar
security vulnerability
## CVE-2020-7020 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>elasticsearch-1.4.4.jar</b></p></summary> <p>Elasticsearch - Open Source, Distributed, RESTful Search Engine</p> <p>Library home page: <a href="http://nexus.sonatype.org/oss-repository-hosting.html/elasticsearch">http://nexus.sonatype.org/oss-repository-hosting.html/elasticsearch</a></p> <p>Path to dependency file: LocalCatalogManager/lcm-packaging/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/elasticsearch/elasticsearch/1.4.4/elasticsearch-1.4.4.jar,/home/wss-scanner/.m2/repository/org/elasticsearch/elasticsearch/1.4.4/elasticsearch-1.4.4.jar</p> <p> Dependency Hierarchy: - MetaModel-full-4.5.5.jar (Root Library) - MetaModel-elasticsearch-rest-4.5.5.jar - :x: **elasticsearch-1.4.4.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Elasticsearch versions before 6.8.13 and 7.9.2 contain a document disclosure flaw when Document or Field Level Security is used. Search queries do not properly preserve security permissions when executing certain complex queries. This could result in the search disclosing the existence of documents the attacker should not be able to view. This could result in an attacker gaining additional insight into potentially sensitive indices. <p>Publish Date: 2020-07-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7020>CVE-2020-7020</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://discuss.elastic.co/t/elastic-stack-7-9-3-and-6-8-13-security-update/253033">https://discuss.elastic.co/t/elastic-stack-7-9-3-and-6-8-13-security-update/253033</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: org.elasticsearch:elasticsearch:6.8.13,7.9.3</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.elasticsearch","packageName":"elasticsearch","packageVersion":"1.4.4","isTransitiveDependency":true,"dependencyTree":"org.apache.metamodel:MetaModel-full:4.5.5;org.apache.metamodel:MetaModel-elasticsearch-rest:4.5.5;org.elasticsearch:elasticsearch:1.4.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.elasticsearch:elasticsearch:6.8.13,7.9.3"}],"vulnerabilityIdentifier":"CVE-2020-7020","vulnerabilityDetails":"Elasticsearch versions before 6.8.13 and 7.9.2 contain a document disclosure flaw when Document or Field Level Security is used. Search queries do not properly preserve security permissions when executing certain complex queries. This could result in the search disclosing the existence of documents the attacker should not be able to view. This could result in an attacker gaining additional insight into potentially sensitive indices.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7020","cvss3Severity":"low","cvss3Score":"3.1","cvss3Metrics":{"A":"None","AC":"High","PR":"Low","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-7020 (Low) detected in elasticsearch-1.4.4.jar - ## CVE-2020-7020 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>elasticsearch-1.4.4.jar</b></p></summary> <p>Elasticsearch - Open Source, Distributed, RESTful Search Engine</p> <p>Library home page: <a href="http://nexus.sonatype.org/oss-repository-hosting.html/elasticsearch">http://nexus.sonatype.org/oss-repository-hosting.html/elasticsearch</a></p> <p>Path to dependency file: LocalCatalogManager/lcm-packaging/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/elasticsearch/elasticsearch/1.4.4/elasticsearch-1.4.4.jar,/home/wss-scanner/.m2/repository/org/elasticsearch/elasticsearch/1.4.4/elasticsearch-1.4.4.jar</p> <p> Dependency Hierarchy: - MetaModel-full-4.5.5.jar (Root Library) - MetaModel-elasticsearch-rest-4.5.5.jar - :x: **elasticsearch-1.4.4.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Elasticsearch versions before 6.8.13 and 7.9.2 contain a document disclosure flaw when Document or Field Level Security is used. Search queries do not properly preserve security permissions when executing certain complex queries. This could result in the search disclosing the existence of documents the attacker should not be able to view. This could result in an attacker gaining additional insight into potentially sensitive indices. <p>Publish Date: 2020-07-21 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7020>CVE-2020-7020</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://discuss.elastic.co/t/elastic-stack-7-9-3-and-6-8-13-security-update/253033">https://discuss.elastic.co/t/elastic-stack-7-9-3-and-6-8-13-security-update/253033</a></p> <p>Release Date: 2020-07-21</p> <p>Fix Resolution: org.elasticsearch:elasticsearch:6.8.13,7.9.3</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.elasticsearch","packageName":"elasticsearch","packageVersion":"1.4.4","isTransitiveDependency":true,"dependencyTree":"org.apache.metamodel:MetaModel-full:4.5.5;org.apache.metamodel:MetaModel-elasticsearch-rest:4.5.5;org.elasticsearch:elasticsearch:1.4.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.elasticsearch:elasticsearch:6.8.13,7.9.3"}],"vulnerabilityIdentifier":"CVE-2020-7020","vulnerabilityDetails":"Elasticsearch versions before 6.8.13 and 7.9.2 contain a document disclosure flaw when Document or Field Level Security is used. Search queries do not properly preserve security permissions when executing certain complex queries. This could result in the search disclosing the existence of documents the attacker should not be able to view. This could result in an attacker gaining additional insight into potentially sensitive indices.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7020","cvss3Severity":"low","cvss3Score":"3.1","cvss3Metrics":{"A":"None","AC":"High","PR":"Low","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_process
cve low detected in elasticsearch jar cve low severity vulnerability vulnerable library elasticsearch jar elasticsearch open source distributed restful search engine library home page a href path to dependency file localcatalogmanager lcm packaging pom xml path to vulnerable library home wss scanner repository org elasticsearch elasticsearch elasticsearch jar home wss scanner repository org elasticsearch elasticsearch elasticsearch jar dependency hierarchy metamodel full jar root library metamodel elasticsearch rest jar x elasticsearch jar vulnerable library found in base branch master vulnerability details elasticsearch versions before and contain a document disclosure flaw when document or field level security is used search queries do not properly preserve security permissions when executing certain complex queries this could result in the search disclosing the existence of documents the attacker should not be able to view this could result in an attacker gaining additional insight into potentially sensitive indices publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org elasticsearch elasticsearch isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails elasticsearch versions before and contain a document disclosure flaw when document or field level security is used search queries do not properly preserve security permissions when executing certain complex queries this could result in the search disclosing the existence of documents the attacker should not be able to view this could result in an attacker gaining additional insight into potentially sensitive indices vulnerabilityurl
0
21,503
29,669,427,460
IssuesEvent
2023-06-11 08:06:32
threefoldtech/tfgrid-sdk-ts
https://api.github.com/repos/threefoldtech/tfgrid-sdk-ts
closed
Proposal: Improve tfchain wallet connector ux
process_wontfix playground
Based on what described in this issue https://github.com/threefoldtech/tfgrid-sdk-ts/issues/488 we are going to allow the user to have only one wallet. This is kind of poor user experience and development experience, let's assume user have 3 *testnet* wallets - 2 *mainnet* wallet; also developers might have 2 dev, 2 qa, 1 test and 1 main. Imagine if u have a list of your wallets whenever you login ![image](https://github.com/threefoldtech/tfgrid-sdk-ts/assets/31689104/3bf982d2-babb-4b2a-acd2-902b3fe7604f) Clicking any one of predefined wallets should show a dialog asking the user to enter it's password ![image](https://github.com/threefoldtech/tfgrid-sdk-ts/assets/31689104/555ee6a5-e7b7-4a99-93af-078d02c667c2) In case the user has no defined wallets he will need to add a new wallet ![image](https://github.com/threefoldtech/tfgrid-sdk-ts/assets/31689104/8c7ca8be-d9c1-46db-a771-9016f3e90a20) ### How this structure can be stored in localstorage? basically we can use a flat pattern by prefixing passwordHash as follow ```yaml localStorage: - wallet.[wallet name].passwordHash: encrypted mnemonic - wallet.devenet wallet1.hash_password_1: xxxxxxxxxxxxxxx1 - wallet.devenet wallet2.hash_password_2: xxxxxxxxxxxxxxx2 ``` ### How to list wallets from localstorage? ```ts for (let i = 0; i < localStorage.length; i++) { const key = localStorage.key(i) // Check if the key is following out structure // if so we should list it } ```
1.0
Proposal: Improve tfchain wallet connector ux - Based on what described in this issue https://github.com/threefoldtech/tfgrid-sdk-ts/issues/488 we are going to allow the user to have only one wallet. This is kind of poor user experience and development experience, let's assume user have 3 *testnet* wallets - 2 *mainnet* wallet; also developers might have 2 dev, 2 qa, 1 test and 1 main. Imagine if u have a list of your wallets whenever you login ![image](https://github.com/threefoldtech/tfgrid-sdk-ts/assets/31689104/3bf982d2-babb-4b2a-acd2-902b3fe7604f) Clicking any one of predefined wallets should show a dialog asking the user to enter it's password ![image](https://github.com/threefoldtech/tfgrid-sdk-ts/assets/31689104/555ee6a5-e7b7-4a99-93af-078d02c667c2) In case the user has no defined wallets he will need to add a new wallet ![image](https://github.com/threefoldtech/tfgrid-sdk-ts/assets/31689104/8c7ca8be-d9c1-46db-a771-9016f3e90a20) ### How this structure can be stored in localstorage? basically we can use a flat pattern by prefixing passwordHash as follow ```yaml localStorage: - wallet.[wallet name].passwordHash: encrypted mnemonic - wallet.devenet wallet1.hash_password_1: xxxxxxxxxxxxxxx1 - wallet.devenet wallet2.hash_password_2: xxxxxxxxxxxxxxx2 ``` ### How to list wallets from localstorage? ```ts for (let i = 0; i < localStorage.length; i++) { const key = localStorage.key(i) // Check if the key is following out structure // if so we should list it } ```
process
proposal improve tfchain wallet connector ux based on what described in this issue we are going to allow the user to have only one wallet this is kind of poor user experience and development experience let s assume user have testnet wallets mainnet wallet also developers might have dev qa test and main imagine if u have a list of your wallets whenever you login clicking any one of predefined wallets should show a dialog asking the user to enter it s password in case the user has no defined wallets he will need to add a new wallet how this structure can be stored in localstorage basically we can use a flat pattern by prefixing passwordhash as follow yaml localstorage wallet passwordhash encrypted mnemonic wallet devenet hash password wallet devenet hash password how to list wallets from localstorage ts for let i i localstorage length i const key localstorage key i check if the key is following out structure if so we should list it
1
10,162
13,044,162,668
IssuesEvent
2020-07-29 03:47:34
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
UCP: Migrate scalar function `RandomBytes` from TiDB
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
## Description Port the scalar function `RandomBytes` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @iosmanthus ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
2.0
UCP: Migrate scalar function `RandomBytes` from TiDB - ## Description Port the scalar function `RandomBytes` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @iosmanthus ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
process
ucp migrate scalar function randombytes from tidb description port the scalar function randombytes from tidb to coprocessor score mentor s iosmanthus recommended skills rust programming learning materials already implemented expressions ported from tidb
1
146,430
5,621,394,047
IssuesEvent
2017-04-04 09:50:29
dangvd/ksmoothdock
https://api.github.com/repos/dangvd/ksmoothdock
opened
Tooltip text is clipped with high DPI settings
Priority: Low Type: Bug
Tooltip's size calculation/drawing needs to handle custom DPI settings
1.0
Tooltip text is clipped with high DPI settings - Tooltip's size calculation/drawing needs to handle custom DPI settings
non_process
tooltip text is clipped with high dpi settings tooltip s size calculation drawing needs to handle custom dpi settings
0
30,898
13,384,055,085
IssuesEvent
2020-09-02 11:22:16
benetech/ServiceNet
https://api.github.com/repos/benetech/ServiceNet
closed
Add 'my location' button to SP map view
ServiceProviderView Story Points: 2 estimated
clicking this button should launch the HTML5 geolocation function and center the map to your location (and refresh results once performance improvements made to map endpoint). https://developers.google.com/maps/documentation/javascript/geolocation Button should be similar to: ![image](https://user-images.githubusercontent.com/12927553/90925293-96191280-e3b6-11ea-832e-b0e1b46aca9d.png)
1.0
Add 'my location' button to SP map view - clicking this button should launch the HTML5 geolocation function and center the map to your location (and refresh results once performance improvements made to map endpoint). https://developers.google.com/maps/documentation/javascript/geolocation Button should be similar to: ![image](https://user-images.githubusercontent.com/12927553/90925293-96191280-e3b6-11ea-832e-b0e1b46aca9d.png)
non_process
add my location button to sp map view clicking this button should launch the geolocation function and center the map to your location and refresh results once performance improvements made to map endpoint button should be similar to
0
12,992
15,358,273,932
IssuesEvent
2021-03-01 14:38:07
prisma/prisma
https://api.github.com/repos/prisma/prisma
closed
Inconsistent Migrate Schema created
process/candidate team/migrations topic: migrate
## Bug description A prisma migration produced alter statements that were inconsistent. Specifically an explicit M:M relationship table was created with both key columns set as NOT NULL, then an ALTER statements created a FK which had an 'ON DELETE SET NULL' clause. During execution, this then failed. ## How to reproduce Sadly, in trying to sort myself out, I tweaked the schema.prisma file, and cannot now reproduce the problem. I believe that I corrected a partial relationship definition in the schema.prisma file, something like: `colname Model @relationship("some_name") ` but I am not really sure and I have been unable to backtrack my modifications and so cannot reproduce the error. ## Expected behavior I now know that I had some issues with my relationship definitions. In particular I had crossed my two required relationships so that addressId referenced the id field of the Person model, and personId referenced the id field of the Address model. That is an error of mine, but was logically correct, and thus the 2 FK Relationships set up were correct although would have given me problems later when I tried to debug problems. I do not expect that Prisma could detect this sort of problem during migration. Then a third FK relationship was constructed by the migration. This FK had an 'ON DELETE SET NULL' clause, despite the table definition setting the column to NOT NULL. This is clearly inconsistent, and should therefore not be possible. At least it should be picked up by Prisma during preparation of the migration SQL, and not at execution time, as in a production system this could leave the application needing a DB restore. Sadly, in trying to sort this out, I have modified the schema.prisma file and can no longer reproduce the error. @tom requested that I submit a bug report nevertheless. As a thought, it would be nice if a) the schema.prisma file was copied to the directory with the corresponding migration.sql as a reference, and b) if it was possible to build the migration.sql file, and then stop the migration at this point so the SQL could be reviewed. Subsequently the migration could be restarted from that point. ## Prisma information <!-- Your Prisma schema, Prisma Client queries, ... **migrate.sql** `/* Warnings: - The migration will add a unique constraint covering the columns `[personId]` on the table `users`. If there are existing duplicate values, the migration will fail. */ -- CreateTable CREATE TABLE `addresses` ( `id` INTEGER NOT NULL AUTO_INCREMENT, `addr1` VARCHAR(191) NOT NULL, `addr2` VARCHAR(191), `addr3` VARCHAR(191), `town` VARCHAR(191), `postCode` VARCHAR(191) NOT NULL, `county` VARCHAR(191), `country` VARCHAR(191), PRIMARY KEY (`id`) ) DEFAULT CHARACTER SET utf8mb4 COLLATE utf8mb4_unicode_ci; -- CreateTable CREATE TABLE `person_address` ( `personId` INTEGER NOT NULL, `addressId` INTEGER NOT NULL, `isMainAddress` BOOLEAN NOT NULL DEFAULT false, INDEX `idx_pa_p`(`personId`), INDEX `idx_pa_a`(`addressId`), PRIMARY KEY (`personId`,`addressId`) ) DEFAULT CHARACTER SET utf8mb4 COLLATE utf8mb4_unicode_ci; -- CreateIndex CREATE UNIQUE INDEX `users_personId_unique` ON `users`(`personId`); -- AddForeignKey 1 ALTER TABLE `person_address` ADD FOREIGN KEY (`personId`) REFERENCES `addresses`(`id`) ON DELETE CASCADE ON UPDATE CASCADE; -- AddForeignKey 2 ALTER TABLE `person_address` ADD FOREIGN KEY (`addressId`) REFERENCES `people`(`id`) ON DELETE CASCADE ON UPDATE CASCADE; -- AddForeignKey 3 ALTER TABLE `person_address` ADD FOREIGN KEY (`addressId`) REFERENCES `addresses`(`id`) ON DELETE SET NULL ON UPDATE CASCADE;` Do not include your database credentials when sharing your Prisma schema! --> ## Environment & setup <!-- In which environment does the problem occur --> - OS: Windows - Database: MySQL - Node.js version: v14.15.0 - Prisma version: 2.17.0 ``` ```
1.0
Inconsistent Migrate Schema created - ## Bug description A prisma migration produced alter statements that were inconsistent. Specifically an explicit M:M relationship table was created with both key columns set as NOT NULL, then an ALTER statements created a FK which had an 'ON DELETE SET NULL' clause. During execution, this then failed. ## How to reproduce Sadly, in trying to sort myself out, I tweaked the schema.prisma file, and cannot now reproduce the problem. I believe that I corrected a partial relationship definition in the schema.prisma file, something like: `colname Model @relationship("some_name") ` but I am not really sure and I have been unable to backtrack my modifications and so cannot reproduce the error. ## Expected behavior I now know that I had some issues with my relationship definitions. In particular I had crossed my two required relationships so that addressId referenced the id field of the Person model, and personId referenced the id field of the Address model. That is an error of mine, but was logically correct, and thus the 2 FK Relationships set up were correct although would have given me problems later when I tried to debug problems. I do not expect that Prisma could detect this sort of problem during migration. Then a third FK relationship was constructed by the migration. This FK had an 'ON DELETE SET NULL' clause, despite the table definition setting the column to NOT NULL. This is clearly inconsistent, and should therefore not be possible. At least it should be picked up by Prisma during preparation of the migration SQL, and not at execution time, as in a production system this could leave the application needing a DB restore. Sadly, in trying to sort this out, I have modified the schema.prisma file and can no longer reproduce the error. @tom requested that I submit a bug report nevertheless. As a thought, it would be nice if a) the schema.prisma file was copied to the directory with the corresponding migration.sql as a reference, and b) if it was possible to build the migration.sql file, and then stop the migration at this point so the SQL could be reviewed. Subsequently the migration could be restarted from that point. ## Prisma information <!-- Your Prisma schema, Prisma Client queries, ... **migrate.sql** `/* Warnings: - The migration will add a unique constraint covering the columns `[personId]` on the table `users`. If there are existing duplicate values, the migration will fail. */ -- CreateTable CREATE TABLE `addresses` ( `id` INTEGER NOT NULL AUTO_INCREMENT, `addr1` VARCHAR(191) NOT NULL, `addr2` VARCHAR(191), `addr3` VARCHAR(191), `town` VARCHAR(191), `postCode` VARCHAR(191) NOT NULL, `county` VARCHAR(191), `country` VARCHAR(191), PRIMARY KEY (`id`) ) DEFAULT CHARACTER SET utf8mb4 COLLATE utf8mb4_unicode_ci; -- CreateTable CREATE TABLE `person_address` ( `personId` INTEGER NOT NULL, `addressId` INTEGER NOT NULL, `isMainAddress` BOOLEAN NOT NULL DEFAULT false, INDEX `idx_pa_p`(`personId`), INDEX `idx_pa_a`(`addressId`), PRIMARY KEY (`personId`,`addressId`) ) DEFAULT CHARACTER SET utf8mb4 COLLATE utf8mb4_unicode_ci; -- CreateIndex CREATE UNIQUE INDEX `users_personId_unique` ON `users`(`personId`); -- AddForeignKey 1 ALTER TABLE `person_address` ADD FOREIGN KEY (`personId`) REFERENCES `addresses`(`id`) ON DELETE CASCADE ON UPDATE CASCADE; -- AddForeignKey 2 ALTER TABLE `person_address` ADD FOREIGN KEY (`addressId`) REFERENCES `people`(`id`) ON DELETE CASCADE ON UPDATE CASCADE; -- AddForeignKey 3 ALTER TABLE `person_address` ADD FOREIGN KEY (`addressId`) REFERENCES `addresses`(`id`) ON DELETE SET NULL ON UPDATE CASCADE;` Do not include your database credentials when sharing your Prisma schema! --> ## Environment & setup <!-- In which environment does the problem occur --> - OS: Windows - Database: MySQL - Node.js version: v14.15.0 - Prisma version: 2.17.0 ``` ```
process
inconsistent migrate schema created bug description a prisma migration produced alter statements that were inconsistent specifically an explicit m m relationship table was created with both key columns set as not null then an alter statements created a fk which had an on delete set null clause during execution this then failed how to reproduce sadly in trying to sort myself out i tweaked the schema prisma file and cannot now reproduce the problem i believe that i corrected a partial relationship definition in the schema prisma file something like colname model relationship some name but i am not really sure and i have been unable to backtrack my modifications and so cannot reproduce the error expected behavior i now know that i had some issues with my relationship definitions in particular i had crossed my two required relationships so that addressid referenced the id field of the person model and personid referenced the id field of the address model that is an error of mine but was logically correct and thus the fk relationships set up were correct although would have given me problems later when i tried to debug problems i do not expect that prisma could detect this sort of problem during migration then a third fk relationship was constructed by the migration this fk had an on delete set null clause despite the table definition setting the column to not null this is clearly inconsistent and should therefore not be possible at least it should be picked up by prisma during preparation of the migration sql and not at execution time as in a production system this could leave the application needing a db restore sadly in trying to sort this out i have modified the schema prisma file and can no longer reproduce the error tom requested that i submit a bug report nevertheless as a thought it would be nice if a the schema prisma file was copied to the directory with the corresponding migration sql as a reference and b if it was possible to build the migration sql file and then stop the migration at this point so the sql could be reviewed subsequently the migration could be restarted from that point prisma information your prisma schema prisma client queries migrate sql warnings the migration will add a unique constraint covering the columns on the table users if there are existing duplicate values the migration will fail createtable create table addresses id integer not null auto increment varchar not null varchar varchar town varchar postcode varchar not null county varchar country varchar primary key id default character set collate unicode ci createtable create table person address personid integer not null addressid integer not null ismainaddress boolean not null default false index idx pa p personid index idx pa a addressid primary key personid addressid default character set collate unicode ci createindex create unique index users personid unique on users personid addforeignkey alter table person address add foreign key personid references addresses id on delete cascade on update cascade addforeignkey alter table person address add foreign key addressid references people id on delete cascade on update cascade addforeignkey alter table person address add foreign key addressid references addresses id on delete set null on update cascade do not include your database credentials when sharing your prisma schema environment setup os windows database mysql node js version prisma version
1
71,576
18,786,169,157
IssuesEvent
2021-11-08 12:22:42
appsmithorg/appsmith
https://api.github.com/repos/appsmithorg/appsmith
closed
[Bug]: List Widget shows cyclic dependency error when children widgets with bindings in action are deleted
Bug App Viewers Pod Critical UI Building Pod Needs Triaging List Widget
### Is there an existing issue for this? - [X] I have searched the existing issues ### Current Behavior Whenever we add binding to the action of the list widget's children and later delete it, the list widget shows cyclic dependencies every time we load the app. ### Steps To Reproduce 1. Drop a list widget 2. Drop an icon button into the list 3. Bind the onClick action of the icon button to `{{List1.backgroundColor}}` and after that delete the icon button. 4. Refresh the page and notice the cyclic dependency error. ### Environment Production ### Version Cloud
1.0
[Bug]: List Widget shows cyclic dependency error when children widgets with bindings in action are deleted - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Current Behavior Whenever we add binding to the action of the list widget's children and later delete it, the list widget shows cyclic dependencies every time we load the app. ### Steps To Reproduce 1. Drop a list widget 2. Drop an icon button into the list 3. Bind the onClick action of the icon button to `{{List1.backgroundColor}}` and after that delete the icon button. 4. Refresh the page and notice the cyclic dependency error. ### Environment Production ### Version Cloud
non_process
list widget shows cyclic dependency error when children widgets with bindings in action are deleted is there an existing issue for this i have searched the existing issues current behavior whenever we add binding to the action of the list widget s children and later delete it the list widget shows cyclic dependencies every time we load the app steps to reproduce drop a list widget drop an icon button into the list bind the onclick action of the icon button to backgroundcolor and after that delete the icon button refresh the page and notice the cyclic dependency error environment production version cloud
0
17,741
23,657,329,425
IssuesEvent
2022-08-26 12:32:50
mdsreq-fga-unb/2022.1-GDS
https://api.github.com/repos/mdsreq-fga-unb/2022.1-GDS
closed
Processo de Desenvolvimento
Processo de Desenvolvimento
**Descrição** a equipe apresenta apenas uma lista de atividades, não organizam elas em um processo.
1.0
Processo de Desenvolvimento - **Descrição** a equipe apresenta apenas uma lista de atividades, não organizam elas em um processo.
process
processo de desenvolvimento descrição a equipe apresenta apenas uma lista de atividades não organizam elas em um processo
1
5,362
8,188,959,651
IssuesEvent
2018-08-30 05:11:07
ryankeefe92/Episodes
https://api.github.com/repos/ryankeefe92/Episodes
closed
For files with multiple audio tracks (usually labeled with "MULTi" tag), have it select the English audio stream and discard the others.
feature preventative process:
A sample multi-track file is in the system Downloads folder for testing
1.0
For files with multiple audio tracks (usually labeled with "MULTi" tag), have it select the English audio stream and discard the others. - A sample multi-track file is in the system Downloads folder for testing
process
for files with multiple audio tracks usually labeled with multi tag have it select the english audio stream and discard the others a sample multi track file is in the system downloads folder for testing
1
12,262
14,787,589,937
IssuesEvent
2021-01-12 07:53:14
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[PM] Site level permission > 'Add Site' button is displayed for users with 'View & Edit' site permission
Bug P2 Participant manager Process: Fixed Process: Tested QA Process: Tested dev
**Stesp:** 1. Assign a user with View and Edit site permission to any closed study 2. Login with that user 3. Navigate to sites tab 4. Observe that 'Add Site' button is displayed **Actual Result:** 'Add Site' button is displayed for users with 'View & Edit' site permission **Expected Result:** : 'Add Site' button should not be displayed for users with 'View & Edit' site permission and instead should be dislayed for Study level and app level Issue observed in QA and Dev ![Screenshot_2](https://user-images.githubusercontent.com/60386291/102987789-e78a0780-4538-11eb-92a4-73a4144e0ccb.png)
3.0
[PM] Site level permission > 'Add Site' button is displayed for users with 'View & Edit' site permission - **Stesp:** 1. Assign a user with View and Edit site permission to any closed study 2. Login with that user 3. Navigate to sites tab 4. Observe that 'Add Site' button is displayed **Actual Result:** 'Add Site' button is displayed for users with 'View & Edit' site permission **Expected Result:** : 'Add Site' button should not be displayed for users with 'View & Edit' site permission and instead should be dislayed for Study level and app level Issue observed in QA and Dev ![Screenshot_2](https://user-images.githubusercontent.com/60386291/102987789-e78a0780-4538-11eb-92a4-73a4144e0ccb.png)
process
site level permission add site button is displayed for users with view edit site permission stesp assign a user with view and edit site permission to any closed study login with that user navigate to sites tab observe that add site button is displayed actual result add site button is displayed for users with view edit site permission expected result add site button should not be displayed for users with view edit site permission and instead should be dislayed for study level and app level issue observed in qa and dev
1
22,123
30,667,555,136
IssuesEvent
2023-07-25 19:31:17
zammad/zammad
https://api.github.com/repos/zammad/zammad
closed
PGP integration
enhancement admin area integration frontend mail processing pgp
### Summary (What) Currently, Zammad is not able to use Pretty Good Privacy (PGP) to sign, encrypt, and decrypt emails. #### Example Use Cases A secure company wants to use PGP to communicate with their customers, because of sensitive content. Also, the usage of S/MIME and PGP together is important for supporting both ways for different outside usage (customers who use S/MIME and some who use PGP). ### Intended Outcome PGP should be usable in Zammad for signing, encrypting, and decrypting emails. On the integration page, an administrator can - see and manage existing keys - trust/untrust keys - delete keys - upload new public and private keys - change the default behavior of signing and encrypting for groups - check the logs for the most recent signing, encrypting, and decrypting operations. In the ticket create/detail view screen, a user can - select if S/MIME or PGP should be used for signing and encrypting. - see in the article meta information if S/MIME or PGP is used. ### How it will work #### PGP Integration The PGP integration needs to be enabled in `System` > `Integrations`. An administrator can perform the following operations in the GUI: - Upload new public keys - Upload new private keys - Remove existing keys In addition to this, the default behavior of signing/encrypting for the existing groups can be specified on a group basis. S/MIME is the more important one, this means if S/MIME and PGP are enabled + usable, we prefer to use S/MIME. Furthermore, to be able to check if signing/encrypting is working fine or to even debug some issues with signing/encrypting, the recent logs can be used. ![Integrations - PGP](https://github.com/zammad/coordination-private-features/assets/6049445/262ed26a-649d-453a-b666-4a3d955a6292) #### Ticket Creation/Detail View The ticket creation and detail view screens are extended to support the PGP integration. During writing a new email, the well-known `Security` field is shown (if PGP and/or S/MIME are enabled). This field is displayed as a switch and it can be decided which security method should be used. Furthermore, it is possible to turn off signing/encrypting. ![Ticket Create](https://github.com/zammad/coordination-private-features/assets/6049445/4957ec49-a7b8-4f2c-a1a3-ed71d51d0f41) The article information in the ticket detail view screen is enriched with information: - Used security method - Encrypted - Signed Errors that happened during signing/encrypting are shown there as well. ![Ticket Zoom](https://github.com/zammad/coordination-private-features/assets/6049445/a0eb2851-8151-4a12-81d6-7121e97a82b0) ### Implementation Details Our PGP implementation is based on [RFC 3156](https://www.rfc-editor.org/rfc/rfc3156). #### PGP REST API - During the upload of a key, we are parsing the key information and save it to the database. - This makes our life easier and removes the need of calling GnuPG every time (performance). - Information to be saved: - Fingerprint - Expiry date - UIDs (mail addresses) - Passphrase (for private keys) - Keys need to be validated during upload. - Duplicates need to be checked. - The used domain needs to be checked. #### In Clarification: Recipient Alias Configuration - We are going to add a new checkbox to the uploading of private keys to determine if a key is used as a `catch-all` key for the domain specified in the key. - This checkbox is controlled by a new hidden setting in Zammad (similar to [Thunderbird](https://support.mozilla.org/de/kb/openpgp-recipient-alias-configuration)). #### Ticket Creation/Detail View - The switch for S/MIME or PGP in the ticket creation/detail view is only shown if S/MIME **and** PGP are enabled. - If S/MIME and PGP are not configured, there is no `Security` section shown at all. - If an administrator does not care about trusted keys, there is a new setting for handling keys as trusted by default. #### Encrypt/Sign - Smart checking for the expiry date of a key - Use case: Key 1 is about to expire in 3 days, a new key is already uploaded - The mechanism to sign/encrypt should detect which key was uploaded last and use it if the identity (mail address) is the same #### Documentation - Public key servers are not supported, please use API to import keys ```[tasklist] ### Tasks - [ ] https://github.com/zammad/coordination-scrum/issues/60 - [ ] https://github.com/zammad/coordination-scrum/issues/59 - [ ] https://github.com/zammad/coordination-scrum/issues/55 - [ ] https://github.com/zammad/coordination-scrum/issues/56 - [ ] https://github.com/zammad/coordination-scrum/issues/57 - [ ] https://github.com/zammad/coordination-scrum/issues/58 - [ ] https://github.com/zammad/coordination-scrum/issues/61 - [ ] https://github.com/zammad/coordination-scrum/issues/63 ```
1.0
PGP integration - ### Summary (What) Currently, Zammad is not able to use Pretty Good Privacy (PGP) to sign, encrypt, and decrypt emails. #### Example Use Cases A secure company wants to use PGP to communicate with their customers, because of sensitive content. Also, the usage of S/MIME and PGP together is important for supporting both ways for different outside usage (customers who use S/MIME and some who use PGP). ### Intended Outcome PGP should be usable in Zammad for signing, encrypting, and decrypting emails. On the integration page, an administrator can - see and manage existing keys - trust/untrust keys - delete keys - upload new public and private keys - change the default behavior of signing and encrypting for groups - check the logs for the most recent signing, encrypting, and decrypting operations. In the ticket create/detail view screen, a user can - select if S/MIME or PGP should be used for signing and encrypting. - see in the article meta information if S/MIME or PGP is used. ### How it will work #### PGP Integration The PGP integration needs to be enabled in `System` > `Integrations`. An administrator can perform the following operations in the GUI: - Upload new public keys - Upload new private keys - Remove existing keys In addition to this, the default behavior of signing/encrypting for the existing groups can be specified on a group basis. S/MIME is the more important one, this means if S/MIME and PGP are enabled + usable, we prefer to use S/MIME. Furthermore, to be able to check if signing/encrypting is working fine or to even debug some issues with signing/encrypting, the recent logs can be used. ![Integrations - PGP](https://github.com/zammad/coordination-private-features/assets/6049445/262ed26a-649d-453a-b666-4a3d955a6292) #### Ticket Creation/Detail View The ticket creation and detail view screens are extended to support the PGP integration. During writing a new email, the well-known `Security` field is shown (if PGP and/or S/MIME are enabled). This field is displayed as a switch and it can be decided which security method should be used. Furthermore, it is possible to turn off signing/encrypting. ![Ticket Create](https://github.com/zammad/coordination-private-features/assets/6049445/4957ec49-a7b8-4f2c-a1a3-ed71d51d0f41) The article information in the ticket detail view screen is enriched with information: - Used security method - Encrypted - Signed Errors that happened during signing/encrypting are shown there as well. ![Ticket Zoom](https://github.com/zammad/coordination-private-features/assets/6049445/a0eb2851-8151-4a12-81d6-7121e97a82b0) ### Implementation Details Our PGP implementation is based on [RFC 3156](https://www.rfc-editor.org/rfc/rfc3156). #### PGP REST API - During the upload of a key, we are parsing the key information and save it to the database. - This makes our life easier and removes the need of calling GnuPG every time (performance). - Information to be saved: - Fingerprint - Expiry date - UIDs (mail addresses) - Passphrase (for private keys) - Keys need to be validated during upload. - Duplicates need to be checked. - The used domain needs to be checked. #### In Clarification: Recipient Alias Configuration - We are going to add a new checkbox to the uploading of private keys to determine if a key is used as a `catch-all` key for the domain specified in the key. - This checkbox is controlled by a new hidden setting in Zammad (similar to [Thunderbird](https://support.mozilla.org/de/kb/openpgp-recipient-alias-configuration)). #### Ticket Creation/Detail View - The switch for S/MIME or PGP in the ticket creation/detail view is only shown if S/MIME **and** PGP are enabled. - If S/MIME and PGP are not configured, there is no `Security` section shown at all. - If an administrator does not care about trusted keys, there is a new setting for handling keys as trusted by default. #### Encrypt/Sign - Smart checking for the expiry date of a key - Use case: Key 1 is about to expire in 3 days, a new key is already uploaded - The mechanism to sign/encrypt should detect which key was uploaded last and use it if the identity (mail address) is the same #### Documentation - Public key servers are not supported, please use API to import keys ```[tasklist] ### Tasks - [ ] https://github.com/zammad/coordination-scrum/issues/60 - [ ] https://github.com/zammad/coordination-scrum/issues/59 - [ ] https://github.com/zammad/coordination-scrum/issues/55 - [ ] https://github.com/zammad/coordination-scrum/issues/56 - [ ] https://github.com/zammad/coordination-scrum/issues/57 - [ ] https://github.com/zammad/coordination-scrum/issues/58 - [ ] https://github.com/zammad/coordination-scrum/issues/61 - [ ] https://github.com/zammad/coordination-scrum/issues/63 ```
process
pgp integration summary what currently zammad is not able to use pretty good privacy pgp to sign encrypt and decrypt emails example use cases a secure company wants to use pgp to communicate with their customers because of sensitive content also the usage of s mime and pgp together is important for supporting both ways for different outside usage customers who use s mime and some who use pgp intended outcome pgp should be usable in zammad for signing encrypting and decrypting emails on the integration page an administrator can see and manage existing keys trust untrust keys delete keys upload new public and private keys change the default behavior of signing and encrypting for groups check the logs for the most recent signing encrypting and decrypting operations in the ticket create detail view screen a user can select if s mime or pgp should be used for signing and encrypting see in the article meta information if s mime or pgp is used how it will work pgp integration the pgp integration needs to be enabled in system integrations an administrator can perform the following operations in the gui upload new public keys upload new private keys remove existing keys in addition to this the default behavior of signing encrypting for the existing groups can be specified on a group basis s mime is the more important one this means if s mime and pgp are enabled usable we prefer to use s mime furthermore to be able to check if signing encrypting is working fine or to even debug some issues with signing encrypting the recent logs can be used ticket creation detail view the ticket creation and detail view screens are extended to support the pgp integration during writing a new email the well known security field is shown if pgp and or s mime are enabled this field is displayed as a switch and it can be decided which security method should be used furthermore it is possible to turn off signing encrypting the article information in the ticket detail view screen is enriched with information used security method encrypted signed errors that happened during signing encrypting are shown there as well implementation details our pgp implementation is based on pgp rest api during the upload of a key we are parsing the key information and save it to the database this makes our life easier and removes the need of calling gnupg every time performance information to be saved fingerprint expiry date uids mail addresses passphrase for private keys keys need to be validated during upload duplicates need to be checked the used domain needs to be checked in clarification recipient alias configuration we are going to add a new checkbox to the uploading of private keys to determine if a key is used as a catch all key for the domain specified in the key this checkbox is controlled by a new hidden setting in zammad similar to ticket creation detail view the switch for s mime or pgp in the ticket creation detail view is only shown if s mime and pgp are enabled if s mime and pgp are not configured there is no security section shown at all if an administrator does not care about trusted keys there is a new setting for handling keys as trusted by default encrypt sign smart checking for the expiry date of a key use case key is about to expire in days a new key is already uploaded the mechanism to sign encrypt should detect which key was uploaded last and use it if the identity mail address is the same documentation public key servers are not supported please use api to import keys tasks
1
17,239
22,961,258,710
IssuesEvent
2022-07-19 15:33:12
hsmusic/hsmusic-wiki
https://api.github.com/repos/hsmusic/hsmusic-wiki
opened
"...by Name" listings should be grouped by A-Z and show "skip to" links at the top
scope: data processing scope: page generation - content thing: listings type: new page / feature
Self-explanatory. Also `#` for numbers and other. We'll probably need to specifically filter *out* everything which doesn't start with A-Z (when normalized) and manually group them together at the end (or start), because locale comparisons may divide them separately relative to A-Z. We could include extra info like "(123 tracks)" in the `<dt>` to pad out the space a little and share that length for fun, 'cuz why not?
1.0
"...by Name" listings should be grouped by A-Z and show "skip to" links at the top - Self-explanatory. Also `#` for numbers and other. We'll probably need to specifically filter *out* everything which doesn't start with A-Z (when normalized) and manually group them together at the end (or start), because locale comparisons may divide them separately relative to A-Z. We could include extra info like "(123 tracks)" in the `<dt>` to pad out the space a little and share that length for fun, 'cuz why not?
process
by name listings should be grouped by a z and show skip to links at the top self explanatory also for numbers and other we ll probably need to specifically filter out everything which doesn t start with a z when normalized and manually group them together at the end or start because locale comparisons may divide them separately relative to a z we could include extra info like tracks in the to pad out the space a little and share that length for fun cuz why not
1
292,213
21,957,176,186
IssuesEvent
2022-05-24 13:06:00
accordproject/cicero
https://api.github.com/repos/accordproject/cicero
closed
Outdated Contributing.md Doc
Type: Documentation 📝
<!--- Provide a general summary of the issue in the Title above --> # Discussion 🗣 <!--- Provide an expanded summary of the issue --> Currently, the contribution guide contains the link to Accord project slack channel. However, Accord project now uses discord instead of slack. So, proper statement update along with link should be done. ## Would you like to work on this issue? Yes
1.0
Outdated Contributing.md Doc - <!--- Provide a general summary of the issue in the Title above --> # Discussion 🗣 <!--- Provide an expanded summary of the issue --> Currently, the contribution guide contains the link to Accord project slack channel. However, Accord project now uses discord instead of slack. So, proper statement update along with link should be done. ## Would you like to work on this issue? Yes
non_process
outdated contributing md doc discussion 🗣 currently the contribution guide contains the link to accord project slack channel however accord project now uses discord instead of slack so proper statement update along with link should be done would you like to work on this issue yes
0
12,757
15,113,564,177
IssuesEvent
2021-02-08 23:57:44
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
[Compile Failed]:Target //src:bazel failed to build
area-EngProd team-XProduct type: support / not a bug (process)
when i compile bazel from binary source,“./compile.sh” I got this. `INFO: From Creating runfiles tree bazel-out/host/bin/third_party/grpc/cpp_plugin.runfiles [for host]: /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/host/bin/third_party/grpc/cpp_plugin.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/host/bin/third_party/grpc/cpp_plugin.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/host/bin/third_party/grpc/cpp_plugin.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/host/bin/third_party/grpc/cpp_plugin.runfiles_manifest: unexpected operator INFO: From Creating runfiles tree bazel-out/local-fastbuild/bin/src/main/java/com/google/devtools/build/lib/bazel/BazelServer.runfiles: /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/main/java/com/google/devtools/build/lib/bazel/BazelServer.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/main/java/com/google/devtools/build/lib/bazel/BazelServer.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/main/java/com/google/devtools/build/lib/bazel/BazelServer.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/main/java/com/google/devtools/build/lib/bazel/BazelServer.runfiles_manifest: unexpected operator INFO: From Compiling src/main/cpp/blaze_util_posix.cc: src/main/cpp/blaze_util_posix.cc: In function 'std::__cxx11::string blaze::RunProgram(const string&, const std::vector<std::__cxx11::basic_string<char> >&)': src/main/cpp/blaze_util_posix.cc:201:1: warning: control reaches end of non-void function [-Wreturn-type] } ^ INFO: From Creating runfiles tree bazel-out/host/bin/third_party/ijar/ijar.runfiles [for host]: /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/host/bin/third_party/ijar/ijar.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/host/bin/third_party/ijar/ijar.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/host/bin/third_party/ijar/ijar.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/host/bin/third_party/ijar/ijar.runfiles_manifest: unexpected operator INFO: From Creating runfiles tree bazel-out/local-fastbuild/bin/third_party/java/apkbuilder/apkbuilder.runfiles: /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/third_party/java/apkbuilder/apkbuilder.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/third_party/java/apkbuilder/apkbuilder.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/third_party/java/apkbuilder/apkbuilder.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/third_party/java/apkbuilder/apkbuilder.runfiles_manifest: unexpected operator INFO: From Creating runfiles tree bazel-out/local-fastbuild/bin/src/java_tools/junitrunner/java/com/google/testing/junit/runner/Runner.runfiles: /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/java_tools/junitrunner/java/com/google/testing/junit/runner/Runner.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/java_tools/junitrunner/java/com/google/testing/junit/runner/Runner.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/java_tools/junitrunner/java/com/google/testing/junit/runner/Runner.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/java_tools/junitrunner/java/com/google/testing/junit/runner/Runner.runfiles_manifest: unexpected operator INFO: From Creating runfiles tree bazel-out/local-fastbuild/bin/src/java_tools/buildjar/JavaBuilder.runfiles: /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/java_tools/buildjar/JavaBuilder.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/java_tools/buildjar/JavaBuilder.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/java_tools/buildjar/JavaBuilder.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/java_tools/buildjar/JavaBuilder.runfiles_manifest: unexpected operator INFO: From Creating runfiles tree bazel-out/local-fastbuild/bin/src/java_tools/singlejar/SingleJar.runfiles: /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/java_tools/singlejar/SingleJar.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/java_tools/singlejar/SingleJar.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/java_tools/singlejar/SingleJar.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/java_tools/singlejar/SingleJar.runfiles_manifest: unexpected operator INFO: From Creating runfiles tree bazel-out/local-fastbuild/bin/src/java_tools/buildjar/java/com/google/devtools/build/java/turbine/turbine.runfiles: /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/java_tools/buildjar/java/com/google/devtools/build/java/turbine/turbine.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/java_tools/buildjar/java/com/google/devtools/build/java/turbine/turbine.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/java_tools/buildjar/java/com/google/devtools/build/java/turbine/turbine.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/java_tools/buildjar/java/com/google/devtools/build/java/turbine/turbine.runfiles_manifest: unexpected operator INFO: From Creating runfiles tree bazel-out/local-fastbuild/bin/src/java_tools/buildjar/java/com/google/devtools/build/buildjar/genclass/GenClass.runfiles: /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/java_tools/buildjar/java/com/google/devtools/build/buildjar/genclass/GenClass.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/java_tools/buildjar/java/com/google/devtools/build/buildjar/genclass/GenClass.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/java_tools/buildjar/java/com/google/devtools/build/buildjar/genclass/GenClass.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/java_tools/buildjar/java/com/google/devtools/build/buildjar/genclass/GenClass.runfiles_manifest: unexpected operator INFO: From Creating runfiles tree bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/idlclass/classes.runfiles: /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/idlclass/classes.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/idlclass/classes.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/idlclass/classes.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/idlclass/classes.runfiles_manifest: unexpected operator INFO: From Creating runfiles tree bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/ziputils/all.runfiles: /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/ziputils/all.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/ziputils/all.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/ziputils/all.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/ziputils/all.runfiles_manifest: unexpected operator INFO: From Creating runfiles tree bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/ideinfo/classes.runfiles: /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/ideinfo/classes.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/ideinfo/classes.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/ideinfo/classes.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/ideinfo/classes.runfiles_manifest: unexpected operator INFO: From Creating runfiles tree bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/classes.runfiles: /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/classes.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/classes.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/classes.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/classes.runfiles_manifest: unexpected operator ERROR: /home/curio/project/bazel/src/tools/android/java/com/google/devtools/build/android/ziputils/BUILD:20:1: Java compilation in rule '//src/tools/android/java/com/google/devtools/build/android/ziputils:ziputils_lib' failed: Worker process did not return a correct WorkResponse. This is probably caused by a bug in the worker, writing unexpected other data to stdout. WARNING: Javac worker failed (java.io.IOException: ?????), invalidating and retrying with new worker... Target //src:bazel failed to build ` Because of JAVA!? HELP ME!
1.0
[Compile Failed]:Target //src:bazel failed to build - when i compile bazel from binary source,“./compile.sh” I got this. `INFO: From Creating runfiles tree bazel-out/host/bin/third_party/grpc/cpp_plugin.runfiles [for host]: /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/host/bin/third_party/grpc/cpp_plugin.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/host/bin/third_party/grpc/cpp_plugin.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/host/bin/third_party/grpc/cpp_plugin.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/host/bin/third_party/grpc/cpp_plugin.runfiles_manifest: unexpected operator INFO: From Creating runfiles tree bazel-out/local-fastbuild/bin/src/main/java/com/google/devtools/build/lib/bazel/BazelServer.runfiles: /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/main/java/com/google/devtools/build/lib/bazel/BazelServer.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/main/java/com/google/devtools/build/lib/bazel/BazelServer.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/main/java/com/google/devtools/build/lib/bazel/BazelServer.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/main/java/com/google/devtools/build/lib/bazel/BazelServer.runfiles_manifest: unexpected operator INFO: From Compiling src/main/cpp/blaze_util_posix.cc: src/main/cpp/blaze_util_posix.cc: In function 'std::__cxx11::string blaze::RunProgram(const string&, const std::vector<std::__cxx11::basic_string<char> >&)': src/main/cpp/blaze_util_posix.cc:201:1: warning: control reaches end of non-void function [-Wreturn-type] } ^ INFO: From Creating runfiles tree bazel-out/host/bin/third_party/ijar/ijar.runfiles [for host]: /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/host/bin/third_party/ijar/ijar.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/host/bin/third_party/ijar/ijar.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/host/bin/third_party/ijar/ijar.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/host/bin/third_party/ijar/ijar.runfiles_manifest: unexpected operator INFO: From Creating runfiles tree bazel-out/local-fastbuild/bin/third_party/java/apkbuilder/apkbuilder.runfiles: /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/third_party/java/apkbuilder/apkbuilder.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/third_party/java/apkbuilder/apkbuilder.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/third_party/java/apkbuilder/apkbuilder.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/third_party/java/apkbuilder/apkbuilder.runfiles_manifest: unexpected operator INFO: From Creating runfiles tree bazel-out/local-fastbuild/bin/src/java_tools/junitrunner/java/com/google/testing/junit/runner/Runner.runfiles: /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/java_tools/junitrunner/java/com/google/testing/junit/runner/Runner.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/java_tools/junitrunner/java/com/google/testing/junit/runner/Runner.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/java_tools/junitrunner/java/com/google/testing/junit/runner/Runner.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/java_tools/junitrunner/java/com/google/testing/junit/runner/Runner.runfiles_manifest: unexpected operator INFO: From Creating runfiles tree bazel-out/local-fastbuild/bin/src/java_tools/buildjar/JavaBuilder.runfiles: /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/java_tools/buildjar/JavaBuilder.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/java_tools/buildjar/JavaBuilder.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/java_tools/buildjar/JavaBuilder.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/java_tools/buildjar/JavaBuilder.runfiles_manifest: unexpected operator INFO: From Creating runfiles tree bazel-out/local-fastbuild/bin/src/java_tools/singlejar/SingleJar.runfiles: /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/java_tools/singlejar/SingleJar.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/java_tools/singlejar/SingleJar.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/java_tools/singlejar/SingleJar.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/java_tools/singlejar/SingleJar.runfiles_manifest: unexpected operator INFO: From Creating runfiles tree bazel-out/local-fastbuild/bin/src/java_tools/buildjar/java/com/google/devtools/build/java/turbine/turbine.runfiles: /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/java_tools/buildjar/java/com/google/devtools/build/java/turbine/turbine.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/java_tools/buildjar/java/com/google/devtools/build/java/turbine/turbine.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/java_tools/buildjar/java/com/google/devtools/build/java/turbine/turbine.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/java_tools/buildjar/java/com/google/devtools/build/java/turbine/turbine.runfiles_manifest: unexpected operator INFO: From Creating runfiles tree bazel-out/local-fastbuild/bin/src/java_tools/buildjar/java/com/google/devtools/build/buildjar/genclass/GenClass.runfiles: /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/java_tools/buildjar/java/com/google/devtools/build/buildjar/genclass/GenClass.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/java_tools/buildjar/java/com/google/devtools/build/buildjar/genclass/GenClass.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/java_tools/buildjar/java/com/google/devtools/build/buildjar/genclass/GenClass.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/java_tools/buildjar/java/com/google/devtools/build/buildjar/genclass/GenClass.runfiles_manifest: unexpected operator INFO: From Creating runfiles tree bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/idlclass/classes.runfiles: /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/idlclass/classes.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/idlclass/classes.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/idlclass/classes.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/idlclass/classes.runfiles_manifest: unexpected operator INFO: From Creating runfiles tree bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/ziputils/all.runfiles: /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/ziputils/all.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/ziputils/all.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/ziputils/all.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/ziputils/all.runfiles_manifest: unexpected operator INFO: From Creating runfiles tree bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/ideinfo/classes.runfiles: /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/ideinfo/classes.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/ideinfo/classes.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/ideinfo/classes.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/ideinfo/classes.runfiles_manifest: unexpected operator INFO: From Creating runfiles tree bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/classes.runfiles: /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/classes.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 4: [: bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/classes.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/classes.runfiles_manifest: unexpected operator /tmp/bazel_lIP1aSJU/out/execroot/bazel/_bin/build-runfiles: 8: [: bazel-out/local-fastbuild/bin/src/tools/android/java/com/google/devtools/build/android/classes.runfiles_manifest: unexpected operator ERROR: /home/curio/project/bazel/src/tools/android/java/com/google/devtools/build/android/ziputils/BUILD:20:1: Java compilation in rule '//src/tools/android/java/com/google/devtools/build/android/ziputils:ziputils_lib' failed: Worker process did not return a correct WorkResponse. This is probably caused by a bug in the worker, writing unexpected other data to stdout. WARNING: Javac worker failed (java.io.IOException: ?????), invalidating and retrying with new worker... Target //src:bazel failed to build ` Because of JAVA!? HELP ME!
process
target src bazel failed to build when i compile bazel from binary source,“ compile sh” i got this info from creating runfiles tree bazel out host bin third party grpc cpp plugin runfiles tmp bazel out execroot bazel bin build runfiles bazel out host bin third party grpc cpp plugin runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out host bin third party grpc cpp plugin runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out host bin third party grpc cpp plugin runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out host bin third party grpc cpp plugin runfiles manifest unexpected operator info from creating runfiles tree bazel out local fastbuild bin src main java com google devtools build lib bazel bazelserver runfiles tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src main java com google devtools build lib bazel bazelserver runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src main java com google devtools build lib bazel bazelserver runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src main java com google devtools build lib bazel bazelserver runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src main java com google devtools build lib bazel bazelserver runfiles manifest unexpected operator info from compiling src main cpp blaze util posix cc src main cpp blaze util posix cc in function std string blaze runprogram const string const std vector src main cpp blaze util posix cc warning control reaches end of non void function info from creating runfiles tree bazel out host bin third party ijar ijar runfiles tmp bazel out execroot bazel bin build runfiles bazel out host bin third party ijar ijar runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out host bin third party ijar ijar runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out host bin third party ijar ijar runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out host bin third party ijar ijar runfiles manifest unexpected operator info from creating runfiles tree bazel out local fastbuild bin third party java apkbuilder apkbuilder runfiles tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin third party java apkbuilder apkbuilder runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin third party java apkbuilder apkbuilder runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin third party java apkbuilder apkbuilder runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin third party java apkbuilder apkbuilder runfiles manifest unexpected operator info from creating runfiles tree bazel out local fastbuild bin src java tools junitrunner java com google testing junit runner runner runfiles tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src java tools junitrunner java com google testing junit runner runner runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src java tools junitrunner java com google testing junit runner runner runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src java tools junitrunner java com google testing junit runner runner runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src java tools junitrunner java com google testing junit runner runner runfiles manifest unexpected operator info from creating runfiles tree bazel out local fastbuild bin src java tools buildjar javabuilder runfiles tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src java tools buildjar javabuilder runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src java tools buildjar javabuilder runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src java tools buildjar javabuilder runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src java tools buildjar javabuilder runfiles manifest unexpected operator info from creating runfiles tree bazel out local fastbuild bin src java tools singlejar singlejar runfiles tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src java tools singlejar singlejar runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src java tools singlejar singlejar runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src java tools singlejar singlejar runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src java tools singlejar singlejar runfiles manifest unexpected operator info from creating runfiles tree bazel out local fastbuild bin src java tools buildjar java com google devtools build java turbine turbine runfiles tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src java tools buildjar java com google devtools build java turbine turbine runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src java tools buildjar java com google devtools build java turbine turbine runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src java tools buildjar java com google devtools build java turbine turbine runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src java tools buildjar java com google devtools build java turbine turbine runfiles manifest unexpected operator info from creating runfiles tree bazel out local fastbuild bin src java tools buildjar java com google devtools build buildjar genclass genclass runfiles tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src java tools buildjar java com google devtools build buildjar genclass genclass runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src java tools buildjar java com google devtools build buildjar genclass genclass runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src java tools buildjar java com google devtools build buildjar genclass genclass runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src java tools buildjar java com google devtools build buildjar genclass genclass runfiles manifest unexpected operator info from creating runfiles tree bazel out local fastbuild bin src tools android java com google devtools build android idlclass classes runfiles tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src tools android java com google devtools build android idlclass classes runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src tools android java com google devtools build android idlclass classes runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src tools android java com google devtools build android idlclass classes runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src tools android java com google devtools build android idlclass classes runfiles manifest unexpected operator info from creating runfiles tree bazel out local fastbuild bin src tools android java com google devtools build android ziputils all runfiles tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src tools android java com google devtools build android ziputils all runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src tools android java com google devtools build android ziputils all runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src tools android java com google devtools build android ziputils all runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src tools android java com google devtools build android ziputils all runfiles manifest unexpected operator info from creating runfiles tree bazel out local fastbuild bin src tools android java com google devtools build android ideinfo classes runfiles tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src tools android java com google devtools build android ideinfo classes runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src tools android java com google devtools build android ideinfo classes runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src tools android java com google devtools build android ideinfo classes runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src tools android java com google devtools build android ideinfo classes runfiles manifest unexpected operator info from creating runfiles tree bazel out local fastbuild bin src tools android java com google devtools build android classes runfiles tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src tools android java com google devtools build android classes runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src tools android java com google devtools build android classes runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src tools android java com google devtools build android classes runfiles manifest unexpected operator tmp bazel out execroot bazel bin build runfiles bazel out local fastbuild bin src tools android java com google devtools build android classes runfiles manifest unexpected operator error home curio project bazel src tools android java com google devtools build android ziputils build java compilation in rule src tools android java com google devtools build android ziputils ziputils lib failed worker process did not return a correct workresponse this is probably caused by a bug in the worker writing unexpected other data to stdout warning javac worker failed java io ioexception invalidating and retrying with new worker target src bazel failed to build because of java help me
1
15,949
20,995,188,782
IssuesEvent
2022-03-29 12:57:51
spring-projects-experimental/spring-native
https://api.github.com/repos/spring-projects-experimental/spring-native
closed
commandlinerunner-log4j2 fails with GraalVM 22.1
type: compatibility
``` openjdk version "17.0.3" 2022-04-19 OpenJDK Runtime Environment GraalVM CE 22.1.0-dev (build 17.0.3+4-jvmci-22.1-b03) OpenJDK 64-Bit Server VM GraalVM CE 22.1.0-dev (build 17.0.3+4-jvmci-22.1-b03, mixed mode, sharing) ``` ``` ======================================================================================================================== GraalVM Native Image: Generating 'commandlinerunner-log4j2' (executable)... ======================================================================================================================== [1/7] Initializing... (0,0s @ 0,07GB) Error: ImageSingletons do not contain key com.oracle.svm.hosted.LinkAtBuildTimeSupport Error: Use -H:+ReportExceptionStackTraces to print stacktrace of underlying exception ------------------------------------------------------------------------------------------------------------------------ 0,1s (6,4% of total time) in 8 GCs | Peak RSS: 0,59GB | CPU load: 5,14 ======================================================================================================================== Failed generating 'commandlinerunner-log4j2' after 1,4s. ``` Looks like https://github.com/spring-projects-experimental/spring-native/issues/1546
True
commandlinerunner-log4j2 fails with GraalVM 22.1 - ``` openjdk version "17.0.3" 2022-04-19 OpenJDK Runtime Environment GraalVM CE 22.1.0-dev (build 17.0.3+4-jvmci-22.1-b03) OpenJDK 64-Bit Server VM GraalVM CE 22.1.0-dev (build 17.0.3+4-jvmci-22.1-b03, mixed mode, sharing) ``` ``` ======================================================================================================================== GraalVM Native Image: Generating 'commandlinerunner-log4j2' (executable)... ======================================================================================================================== [1/7] Initializing... (0,0s @ 0,07GB) Error: ImageSingletons do not contain key com.oracle.svm.hosted.LinkAtBuildTimeSupport Error: Use -H:+ReportExceptionStackTraces to print stacktrace of underlying exception ------------------------------------------------------------------------------------------------------------------------ 0,1s (6,4% of total time) in 8 GCs | Peak RSS: 0,59GB | CPU load: 5,14 ======================================================================================================================== Failed generating 'commandlinerunner-log4j2' after 1,4s. ``` Looks like https://github.com/spring-projects-experimental/spring-native/issues/1546
non_process
commandlinerunner fails with graalvm openjdk version openjdk runtime environment graalvm ce dev build jvmci openjdk bit server vm graalvm ce dev build jvmci mixed mode sharing graalvm native image generating commandlinerunner executable initializing error imagesingletons do not contain key com oracle svm hosted linkatbuildtimesupport error use h reportexceptionstacktraces to print stacktrace of underlying exception of total time in gcs peak rss cpu load failed generating commandlinerunner after looks like
0
20,182
26,744,656,450
IssuesEvent
2023-01-30 15:14:21
TUM-Dev/NavigaTUM
https://api.github.com/repos/TUM-Dev/NavigaTUM
opened
[Entry] [5204]: Koordinate bearbeiten
entry webform delete-after-processing
Hallo, ich möchte diese Koordinate im Roomfinder korrigieren: ```yaml "5204": { lat: 48.26480782182901, lon: 11.672907360363183 } ```
1.0
[Entry] [5204]: Koordinate bearbeiten - Hallo, ich möchte diese Koordinate im Roomfinder korrigieren: ```yaml "5204": { lat: 48.26480782182901, lon: 11.672907360363183 } ```
process
koordinate bearbeiten hallo ich möchte diese koordinate im roomfinder korrigieren yaml lat lon
1
9,674
12,678,095,947
IssuesEvent
2020-06-19 09:08:29
prisma/prisma-engines
https://api.github.com/repos/prisma/prisma-engines
opened
Check whether the ForeignKeyDefaultDrop destructive change check is still needed
component: migration engine process/candidate
This is an old destructive change warning, may not be needed anymore with the new relations syntax.
1.0
Check whether the ForeignKeyDefaultDrop destructive change check is still needed - This is an old destructive change warning, may not be needed anymore with the new relations syntax.
process
check whether the foreignkeydefaultdrop destructive change check is still needed this is an old destructive change warning may not be needed anymore with the new relations syntax
1
28,787
23,493,675,520
IssuesEvent
2022-08-17 21:34:12
gazebosim/docs
https://api.github.com/repos/gazebosim/docs
closed
Organize docs by major version, not minor
infrastructure
Currently, library docs are organized by their minor version. For example, this URL is specific to version 6.4, not to version 6: https://ignitionrobotics.org/api/math/6.4/index.html Since we don't break behaviour within a major version, and we often make minor releases, I think it would be more convenient to just separate documentation according to major version, instead of minor. If needed, a tutorial can mention from which minor version the feature is available. This way, in the case above, when we release ign-math 6.5, we don't need to release a whole new set of docs.
1.0
Organize docs by major version, not minor - Currently, library docs are organized by their minor version. For example, this URL is specific to version 6.4, not to version 6: https://ignitionrobotics.org/api/math/6.4/index.html Since we don't break behaviour within a major version, and we often make minor releases, I think it would be more convenient to just separate documentation according to major version, instead of minor. If needed, a tutorial can mention from which minor version the feature is available. This way, in the case above, when we release ign-math 6.5, we don't need to release a whole new set of docs.
non_process
organize docs by major version not minor currently library docs are organized by their minor version for example this url is specific to version not to version since we don t break behaviour within a major version and we often make minor releases i think it would be more convenient to just separate documentation according to major version instead of minor if needed a tutorial can mention from which minor version the feature is available this way in the case above when we release ign math we don t need to release a whole new set of docs
0
3,329
6,447,523,043
IssuesEvent
2017-08-14 07:44:30
openvstorage/integrationtests
https://api.github.com/repos/openvstorage/integrationtests
closed
VM creation fails on EE-VIR-4N-199.191-fwk-develop in ci_scenario_edge_test
process_duplicate
Environment EE-VIR-4N-199.191-fwk-develop ``` 2017-07-26 00:50:05 99600 +0200 - ovs-node01-1604 - 11552/140699579442944 - scenario_helpers/vm_handler - 3346 - INFO - Boot VDisk successfully created. 2017-07-26 00:50:17 65200 +0200 - ovs-node01-1604 - 11552/140699579442944 - scenario_helpers/vm_handler - 3360 - INFO - VDisk data_vdisk successfully created! 2017-07-26 00:50:18 11300 +0200 - ovs-node01-1604 - 11552/140699579442944 - scenario_helpers/vm_handler - 3362 - WARNING - Could not fetch the cd vdisk after 0.0208010673523s. 2017-07-26 00:50:18 61600 +0200 - ovs-node01-1604 - 11552/140699579442944 - scenario_helpers/vm_handler - 3363 - WARNING - Could not fetch the cd vdisk after 0.523992061615s. 2017-07-26 00:50:19 11900 +0200 - ovs-node01-1604 - 11552/140699579442944 - scenario_helpers/vm_handler - 3364 - WARNING - Could not fetch the cd vdisk after 1.02754616737s. 2017-07-26 00:50:20 25700 +0200 - ovs-node01-1604 - 11552/140699579442944 - scenario_helpers/vm_handler - 3367 - INFO - Prepped everything for VM mds-regression-000. 2017-07-26 00:50:20 26800 +0200 - ovs-node01-1604 - 11552/140699579442944 - scenario_helpers/vm_handler - 3369 - INFO - Creating VM `mds-regression-000` 2017-07-26 00:50:21 94000 +0200 - ovs-node01-1604 - 11552/140698181281536 - scenario_helpers/vm_handler - 3372 - ERROR - Error while listening for VM messages.. Got [Errno 98] Address already in use: None ========= Remote Traceback (1) ========= Traceback (most recent call last): File "/tmp/tmp.CSo0D3Vz7U/rpyc/core/protocol.py", line 305, in _dispatch_request res = self._HANDLERS[handler](self, *args) File "/tmp/tmp.CSo0D3Vz7U/rpyc/core/protocol.py", line 535, in _handle_call return self._local_objects[oid](*args, **dict(kwargs)) File "/usr/lib/python2.7/socket.py", line 228, in meth return getattr(self._sock,name)(*args) error: [Errno 98] Address already in use 2017-07-26 00:50:23 90300 +0200 - ovs-node01-1604 - 11552/140699579442944 - scenario_helpers/vm_handler - 3377 - INFO - Created VM `mds-regression-000`! ```
1.0
VM creation fails on EE-VIR-4N-199.191-fwk-develop in ci_scenario_edge_test - Environment EE-VIR-4N-199.191-fwk-develop ``` 2017-07-26 00:50:05 99600 +0200 - ovs-node01-1604 - 11552/140699579442944 - scenario_helpers/vm_handler - 3346 - INFO - Boot VDisk successfully created. 2017-07-26 00:50:17 65200 +0200 - ovs-node01-1604 - 11552/140699579442944 - scenario_helpers/vm_handler - 3360 - INFO - VDisk data_vdisk successfully created! 2017-07-26 00:50:18 11300 +0200 - ovs-node01-1604 - 11552/140699579442944 - scenario_helpers/vm_handler - 3362 - WARNING - Could not fetch the cd vdisk after 0.0208010673523s. 2017-07-26 00:50:18 61600 +0200 - ovs-node01-1604 - 11552/140699579442944 - scenario_helpers/vm_handler - 3363 - WARNING - Could not fetch the cd vdisk after 0.523992061615s. 2017-07-26 00:50:19 11900 +0200 - ovs-node01-1604 - 11552/140699579442944 - scenario_helpers/vm_handler - 3364 - WARNING - Could not fetch the cd vdisk after 1.02754616737s. 2017-07-26 00:50:20 25700 +0200 - ovs-node01-1604 - 11552/140699579442944 - scenario_helpers/vm_handler - 3367 - INFO - Prepped everything for VM mds-regression-000. 2017-07-26 00:50:20 26800 +0200 - ovs-node01-1604 - 11552/140699579442944 - scenario_helpers/vm_handler - 3369 - INFO - Creating VM `mds-regression-000` 2017-07-26 00:50:21 94000 +0200 - ovs-node01-1604 - 11552/140698181281536 - scenario_helpers/vm_handler - 3372 - ERROR - Error while listening for VM messages.. Got [Errno 98] Address already in use: None ========= Remote Traceback (1) ========= Traceback (most recent call last): File "/tmp/tmp.CSo0D3Vz7U/rpyc/core/protocol.py", line 305, in _dispatch_request res = self._HANDLERS[handler](self, *args) File "/tmp/tmp.CSo0D3Vz7U/rpyc/core/protocol.py", line 535, in _handle_call return self._local_objects[oid](*args, **dict(kwargs)) File "/usr/lib/python2.7/socket.py", line 228, in meth return getattr(self._sock,name)(*args) error: [Errno 98] Address already in use 2017-07-26 00:50:23 90300 +0200 - ovs-node01-1604 - 11552/140699579442944 - scenario_helpers/vm_handler - 3377 - INFO - Created VM `mds-regression-000`! ```
process
vm creation fails on ee vir fwk develop in ci scenario edge test environment ee vir fwk develop ovs scenario helpers vm handler info boot vdisk successfully created ovs scenario helpers vm handler info vdisk data vdisk successfully created ovs scenario helpers vm handler warning could not fetch the cd vdisk after ovs scenario helpers vm handler warning could not fetch the cd vdisk after ovs scenario helpers vm handler warning could not fetch the cd vdisk after ovs scenario helpers vm handler info prepped everything for vm mds regression ovs scenario helpers vm handler info creating vm mds regression ovs scenario helpers vm handler error error while listening for vm messages got address already in use none remote traceback traceback most recent call last file tmp tmp rpyc core protocol py line in dispatch request res self handlers self args file tmp tmp rpyc core protocol py line in handle call return self local objects args dict kwargs file usr lib socket py line in meth return getattr self sock name args error address already in use ovs scenario helpers vm handler info created vm mds regression
1
7,675
10,761,255,866
IssuesEvent
2019-10-31 20:22:50
googleapis/google-cloud-python
https://api.github.com/repos/googleapis/google-cloud-python
closed
PubSub: Add system test for streaming pull using an account with pubsub.subscriber role
api: pubsub testing type: process
A PubSub regression (#9339) was introduced in the 1.0.1, and fixed in the `1.0.2` release. The fix should be covered by a proper system test. The test should run a simple streaming pull scenario using a service account that only has the `pubsub.subscriber` role, and no other additional permissions. This implies creating a service account, and adding it to the Kokoro environment (contact @busunkim96 for more info). **Side question:** How to easily share the same account(s) with developers that need/want to run system tests locally? **Resources:** - [PubSub roles](https://cloud.google.com/pubsub/docs/access-control#roles)
1.0
PubSub: Add system test for streaming pull using an account with pubsub.subscriber role - A PubSub regression (#9339) was introduced in the 1.0.1, and fixed in the `1.0.2` release. The fix should be covered by a proper system test. The test should run a simple streaming pull scenario using a service account that only has the `pubsub.subscriber` role, and no other additional permissions. This implies creating a service account, and adding it to the Kokoro environment (contact @busunkim96 for more info). **Side question:** How to easily share the same account(s) with developers that need/want to run system tests locally? **Resources:** - [PubSub roles](https://cloud.google.com/pubsub/docs/access-control#roles)
process
pubsub add system test for streaming pull using an account with pubsub subscriber role a pubsub regression was introduced in the and fixed in the release the fix should be covered by a proper system test the test should run a simple streaming pull scenario using a service account that only has the pubsub subscriber role and no other additional permissions this implies creating a service account and adding it to the kokoro environment contact for more info side question how to easily share the same account s with developers that need want to run system tests locally resources
1
28
2,497,037,177
IssuesEvent
2015-01-07 00:31:37
tinkerpop/tinkerpop3
https://api.github.com/repos/tinkerpop/tinkerpop3
opened
Go through all lambda steps and remove Traverser<S> and replace with <S>
enhancement process
The only steps that will support `Lambda<Traverser<S>>` will be `map()`, `flatMap()`, `filter()`, `sideEffect()`, `branch()`, and `repeat()`. All of those are generic lambda steps (besides `repeat()`). `repeat()` requires `Traverser<S>` so that loops can be checked. Note that this has already been updated for most steps with `by()`-modulation. `groupCount().by('name')`. The only danglers that I can think of are `choose()`, `union()`, `fold()`, .. ??... This should make @mpollmeier happy as this was one of his suggestions back in M1 days.
1.0
Go through all lambda steps and remove Traverser<S> and replace with <S> - The only steps that will support `Lambda<Traverser<S>>` will be `map()`, `flatMap()`, `filter()`, `sideEffect()`, `branch()`, and `repeat()`. All of those are generic lambda steps (besides `repeat()`). `repeat()` requires `Traverser<S>` so that loops can be checked. Note that this has already been updated for most steps with `by()`-modulation. `groupCount().by('name')`. The only danglers that I can think of are `choose()`, `union()`, `fold()`, .. ??... This should make @mpollmeier happy as this was one of his suggestions back in M1 days.
process
go through all lambda steps and remove traverser and replace with the only steps that will support lambda will be map flatmap filter sideeffect branch and repeat all of those are generic lambda steps besides repeat repeat requires traverser so that loops can be checked note that this has already been updated for most steps with by modulation groupcount by name the only danglers that i can think of are choose union fold this should make mpollmeier happy as this was one of his suggestions back in days
1
3,185
6,259,016,860
IssuesEvent
2017-07-14 17:00:29
PeaceGeeksSociety/salesforce
https://api.github.com/repos/PeaceGeeksSociety/salesforce
opened
Store interview notes with contact/lead
Recruitment Processes
We would like to define a process that allows us to upload notes onto contact/lead as an attachment. This is because it would allow us to easily identify worthwhile candidates that we can read out to if any positions open up. Done when: process defined for uploading notes onto contact/lead as an attachment, all previous interview notes have been added to contacts.
1.0
Store interview notes with contact/lead - We would like to define a process that allows us to upload notes onto contact/lead as an attachment. This is because it would allow us to easily identify worthwhile candidates that we can read out to if any positions open up. Done when: process defined for uploading notes onto contact/lead as an attachment, all previous interview notes have been added to contacts.
process
store interview notes with contact lead we would like to define a process that allows us to upload notes onto contact lead as an attachment this is because it would allow us to easily identify worthwhile candidates that we can read out to if any positions open up done when process defined for uploading notes onto contact lead as an attachment all previous interview notes have been added to contacts
1
81,598
23,507,810,527
IssuesEvent
2022-08-18 14:01:57
haskell/cabal
https://api.github.com/repos/haskell/cabal
closed
cabal file not found despite --cabal-file
cabal-install: nix-local-build resolution: wontfix
**Describe the bug** I expect to be able to use cabal files at an abnormal path by pointing Cabal at it with the `--cabal-file` argument. **To Reproduce** Steps to reproduce the behavior: 1. Set up a package that builds normally with Cabal. 2. Move `[package name].cabal` outside of the directory or change the extension. 3. Run `cabal v2-build --cabal-file [new path]`. **Expected behavior** Cabal builds the package as if the Cabal file was at its original path. **Actual behavior** Cabal reports the following: ``` No cabal.project file or cabal file matching the default glob './*.cabal' was found. Please create a package description file <pkgname>.cabal or a cabal.project file referencing the packages you want to build. ``` **System information** - Operating system: NixOS 20.03 - `cabal` version: 3.2.0.0
1.0
cabal file not found despite --cabal-file - **Describe the bug** I expect to be able to use cabal files at an abnormal path by pointing Cabal at it with the `--cabal-file` argument. **To Reproduce** Steps to reproduce the behavior: 1. Set up a package that builds normally with Cabal. 2. Move `[package name].cabal` outside of the directory or change the extension. 3. Run `cabal v2-build --cabal-file [new path]`. **Expected behavior** Cabal builds the package as if the Cabal file was at its original path. **Actual behavior** Cabal reports the following: ``` No cabal.project file or cabal file matching the default glob './*.cabal' was found. Please create a package description file <pkgname>.cabal or a cabal.project file referencing the packages you want to build. ``` **System information** - Operating system: NixOS 20.03 - `cabal` version: 3.2.0.0
non_process
cabal file not found despite cabal file describe the bug i expect to be able to use cabal files at an abnormal path by pointing cabal at it with the cabal file argument to reproduce steps to reproduce the behavior set up a package that builds normally with cabal move cabal outside of the directory or change the extension run cabal build cabal file expected behavior cabal builds the package as if the cabal file was at its original path actual behavior cabal reports the following no cabal project file or cabal file matching the default glob cabal was found please create a package description file cabal or a cabal project file referencing the packages you want to build system information operating system nixos cabal version
0
107,305
9,206,210,133
IssuesEvent
2019-03-08 13:04:35
skaut/SRS
https://api.github.com/repos/skaut/SRS
closed
Chyba: při pokusu vstoupit na stránku bez oprávnění
Testování
Pokud se pokusím vstoupit na stránku, na kterou nemám oprávnění zobrazí se chyba: například: https://pocin.skauting.cz/page/kde-kdy-jak/default Zobrazí: "Oops... Your browser sent a request that this server could not understand or process." Pokud vstupuji na stránku, na kterou nemám práva, tak by bylo nejlepší zobrazit stránku s varováním: "K zobrazení této stránky nemáte práva" Děkuji
1.0
Chyba: při pokusu vstoupit na stránku bez oprávnění - Pokud se pokusím vstoupit na stránku, na kterou nemám oprávnění zobrazí se chyba: například: https://pocin.skauting.cz/page/kde-kdy-jak/default Zobrazí: "Oops... Your browser sent a request that this server could not understand or process." Pokud vstupuji na stránku, na kterou nemám práva, tak by bylo nejlepší zobrazit stránku s varováním: "K zobrazení této stránky nemáte práva" Děkuji
non_process
chyba při pokusu vstoupit na stránku bez oprávnění pokud se pokusím vstoupit na stránku na kterou nemám oprávnění zobrazí se chyba například zobrazí oops your browser sent a request that this server could not understand or process pokud vstupuji na stránku na kterou nemám práva tak by bylo nejlepší zobrazit stránku s varováním k zobrazení této stránky nemáte práva děkuji
0
11,335
14,147,291,872
IssuesEvent
2020-11-10 20:33:57
retaildevcrews/ngsa
https://api.github.com/repos/retaildevcrews/ngsa
closed
NGSA - Retro - M1 - Sprint1
Process Retro
## NGSA - Retro - M1 - Sprint1 ### Link: https://rdc-retro.azurewebsites.net/game/-CwUMvXs6 ### Summary **Date**: Tuesday, 10 November 2020 at 14:32:57 GMT-06:00 #### Liked - (+13/-0) The collaboration of the team - a great group to work with! - (+9/-0) We have logs flowing and have already gathered some insights and have some hypothesis. - (+7/-0) Being able to leverage a lot of the NGSA work for a successful One Week project - (+7/-0) Seeing impromptu collab meetings popping up - (+5/-0) Lots of progress across all epics. Very impactful sprint. #### Learned - (+6/-0) The importance of having kubernetes aliases for common tasks :) - (+5/-0) A lot more about Log Analytics - (+5/-0) Multipass :) - (+4/-0) A bit about DestinationRules, subsets, routes, virtualservices, etc. - (+3/-0) How to build and deploy a sidecar. - (+3/-0) Got to learn about Istio and how some of the components work. Also learned that there is A LOT more to learn about Istio and service meshes! - (+3/-0) Grafana dashboard with Azure Monitor - (+2/-0) Power BI has a new look/layout and moved things around, but lost some features - (+2/-0) Learned how to use credscan. Shout out to Bart for walkthrough. - (+2/-0) Learned that resource groups can have resources in different locations. - (+2/-0) SSL/TLS certs, enabling Https for a web service and cert manager - (+0/-0) Learned that you can have k8s health probes for non http services #### Lacked - (+4/-0) More time for One "Week" #### Longed For - (+6/-0) Turf and surf; 2nd Bar; Lonesome Dove; Cooper's; Taco Deli; SPiN - WITH the team :( - (+0/-3) ^- Does this mean someone went to SPIN without the team? 😮 #### Thanks to/for - (+7/-0) The whole team for jumping in on short notice repeatedly to unblock teammates. - (+5/-0) Anne for her pirate skills and creating a picture that showed exactly what I was trying to say. aRRRRRRRRRRRRRRRRR - (+5/-0) Kevin joining the team - (+4/-0) A lot of pairing/group sessions with the team. - (+1/-0) Bart for great documentation in the BM setup, made it very easy to mirror functionality in AKS setup ### Action Items ### Tasks - [X] Add Retro summary into Issue - [ ] Create issues for relevant action items - [ ] If final sprint for milestone. create milestone summary to share
1.0
NGSA - Retro - M1 - Sprint1 - ## NGSA - Retro - M1 - Sprint1 ### Link: https://rdc-retro.azurewebsites.net/game/-CwUMvXs6 ### Summary **Date**: Tuesday, 10 November 2020 at 14:32:57 GMT-06:00 #### Liked - (+13/-0) The collaboration of the team - a great group to work with! - (+9/-0) We have logs flowing and have already gathered some insights and have some hypothesis. - (+7/-0) Being able to leverage a lot of the NGSA work for a successful One Week project - (+7/-0) Seeing impromptu collab meetings popping up - (+5/-0) Lots of progress across all epics. Very impactful sprint. #### Learned - (+6/-0) The importance of having kubernetes aliases for common tasks :) - (+5/-0) A lot more about Log Analytics - (+5/-0) Multipass :) - (+4/-0) A bit about DestinationRules, subsets, routes, virtualservices, etc. - (+3/-0) How to build and deploy a sidecar. - (+3/-0) Got to learn about Istio and how some of the components work. Also learned that there is A LOT more to learn about Istio and service meshes! - (+3/-0) Grafana dashboard with Azure Monitor - (+2/-0) Power BI has a new look/layout and moved things around, but lost some features - (+2/-0) Learned how to use credscan. Shout out to Bart for walkthrough. - (+2/-0) Learned that resource groups can have resources in different locations. - (+2/-0) SSL/TLS certs, enabling Https for a web service and cert manager - (+0/-0) Learned that you can have k8s health probes for non http services #### Lacked - (+4/-0) More time for One "Week" #### Longed For - (+6/-0) Turf and surf; 2nd Bar; Lonesome Dove; Cooper's; Taco Deli; SPiN - WITH the team :( - (+0/-3) ^- Does this mean someone went to SPIN without the team? 😮 #### Thanks to/for - (+7/-0) The whole team for jumping in on short notice repeatedly to unblock teammates. - (+5/-0) Anne for her pirate skills and creating a picture that showed exactly what I was trying to say. aRRRRRRRRRRRRRRRRR - (+5/-0) Kevin joining the team - (+4/-0) A lot of pairing/group sessions with the team. - (+1/-0) Bart for great documentation in the BM setup, made it very easy to mirror functionality in AKS setup ### Action Items ### Tasks - [X] Add Retro summary into Issue - [ ] Create issues for relevant action items - [ ] If final sprint for milestone. create milestone summary to share
process
ngsa retro ngsa retro link summary date tuesday november at gmt liked the collaboration of the team a great group to work with we have logs flowing and have already gathered some insights and have some hypothesis being able to leverage a lot of the ngsa work for a successful one week project seeing impromptu collab meetings popping up lots of progress across all epics very impactful sprint learned the importance of having kubernetes aliases for common tasks a lot more about log analytics multipass a bit about destinationrules subsets routes virtualservices etc how to build and deploy a sidecar got to learn about istio and how some of the components work also learned that there is a lot more to learn about istio and service meshes grafana dashboard with azure monitor power bi has a new look layout and moved things around but lost some features learned how to use credscan shout out to bart for walkthrough learned that resource groups can have resources in different locations ssl tls certs enabling https for a web service and cert manager learned that you can have health probes for non http services lacked more time for one week longed for turf and surf bar lonesome dove cooper s taco deli spin with the team does this mean someone went to spin without the team 😮 thanks to for the whole team for jumping in on short notice repeatedly to unblock teammates anne for her pirate skills and creating a picture that showed exactly what i was trying to say arrrrrrrrrrrrrrrrr kevin joining the team a lot of pairing group sessions with the team bart for great documentation in the bm setup made it very easy to mirror functionality in aks setup action items tasks add retro summary into issue create issues for relevant action items if final sprint for milestone create milestone summary to share
1
16,259
20,827,338,404
IssuesEvent
2022-03-18 23:21:17
googleapis/retry-request
https://api.github.com/repos/googleapis/retry-request
opened
add linting build
type: process priority: p2 api: N/A
I'm adding the standard ci workflow, but couldn't make the lint build pass. We should add lint build.
1.0
add linting build - I'm adding the standard ci workflow, but couldn't make the lint build pass. We should add lint build.
process
add linting build i m adding the standard ci workflow but couldn t make the lint build pass we should add lint build
1
140,488
5,410,388,485
IssuesEvent
2017-03-01 08:33:42
pufexi/multiorder
https://api.github.com/repos/pufexi/multiorder
closed
Implementace EET do generovani faktur v PDF
high priority
Vice: https://github.com/slevomat/eet-client 1) Predstava je takova, ze se bude tisknout ten kod do PDF, ktere se generuje. 2) Dale se bude ten kod ukladat ke kazde objednavce, tj. do tabulky orders
1.0
Implementace EET do generovani faktur v PDF - Vice: https://github.com/slevomat/eet-client 1) Predstava je takova, ze se bude tisknout ten kod do PDF, ktere se generuje. 2) Dale se bude ten kod ukladat ke kazde objednavce, tj. do tabulky orders
non_process
implementace eet do generovani faktur v pdf vice predstava je takova ze se bude tisknout ten kod do pdf ktere se generuje dale se bude ten kod ukladat ke kazde objednavce tj do tabulky orders
0
16,190
20,628,503,183
IssuesEvent
2022-03-08 02:32:30
cypress-io/cypress
https://api.github.com/repos/cypress-io/cypress
closed
Document patch-package
process: tests process: contributing stage: needs review
Add contributor docs around: 1. our use of patch-package 2. testing policy for `patch-package` (either have a `binary-system-test` covering the patched behavior or add the patch to `testStaticAssets.js` Most likely in `guides/patch-package.md`.
2.0
Document patch-package - Add contributor docs around: 1. our use of patch-package 2. testing policy for `patch-package` (either have a `binary-system-test` covering the patched behavior or add the patch to `testStaticAssets.js` Most likely in `guides/patch-package.md`.
process
document patch package add contributor docs around our use of patch package testing policy for patch package either have a binary system test covering the patched behavior or add the patch to teststaticassets js most likely in guides patch package md
1
180,212
14,742,137,184
IssuesEvent
2021-01-07 11:46:03
avinashkranjan/Amazing-Python-Scripts
https://api.github.com/repos/avinashkranjan/Amazing-Python-Scripts
closed
Missing .gitignore?
SWoC 2021 assigned documentation easy enhancement good first issue
**Is your feature request related to a problem? Please describe.** So apparently we are missing a `.gitignore`. That is not a major problem obviously, but I do believe that if say I was a newcomer, I would have a hard time pushing my code without including the say `.idea` or `.vs` or `jupyter-checkpoints` **Describe the solution you'd like** A `.gitignore` **Describe alternatives you've considered** The simplest solution is just making sure that IDE generated files are not committed, it is effective, Yes. But Not optimized. **Additional context** In case there is a particular reason why we are missing `.gtignore` I am sorry as I want aware of that
1.0
Missing .gitignore? - **Is your feature request related to a problem? Please describe.** So apparently we are missing a `.gitignore`. That is not a major problem obviously, but I do believe that if say I was a newcomer, I would have a hard time pushing my code without including the say `.idea` or `.vs` or `jupyter-checkpoints` **Describe the solution you'd like** A `.gitignore` **Describe alternatives you've considered** The simplest solution is just making sure that IDE generated files are not committed, it is effective, Yes. But Not optimized. **Additional context** In case there is a particular reason why we are missing `.gtignore` I am sorry as I want aware of that
non_process
missing gitignore is your feature request related to a problem please describe so apparently we are missing a gitignore that is not a major problem obviously but i do believe that if say i was a newcomer i would have a hard time pushing my code without including the say idea or vs or jupyter checkpoints describe the solution you d like a gitignore describe alternatives you ve considered the simplest solution is just making sure that ide generated files are not committed it is effective yes but not optimized additional context in case there is a particular reason why we are missing gtignore i am sorry as i want aware of that
0
12,442
14,933,538,391
IssuesEvent
2021-01-25 09:21:08
jgraley/inferno-cpp2v
https://api.github.com/repos/jgraley/inferno-cpp2v
opened
deduce variables in one way only
Constraint Processing
`SimpleSolver::Plan::DeduceVariables` can work in two different ways. Pick one and commit to it. Code can always be changed later.
1.0
deduce variables in one way only - `SimpleSolver::Plan::DeduceVariables` can work in two different ways. Pick one and commit to it. Code can always be changed later.
process
deduce variables in one way only simplesolver plan deducevariables can work in two different ways pick one and commit to it code can always be changed later
1
11,066
13,901,583,940
IssuesEvent
2020-10-20 03:16:09
kubeflow/community
https://api.github.com/repos/kubeflow/community
closed
Missing GitHub group kubeflow/wg-notebooks-leads
area/community kind/process priority/p0
@kimwnasptd @StefanoFioravanzo @elikatsis @thesuperzapper Could one of you create the PR to add the group please?
1.0
Missing GitHub group kubeflow/wg-notebooks-leads - @kimwnasptd @StefanoFioravanzo @elikatsis @thesuperzapper Could one of you create the PR to add the group please?
process
missing github group kubeflow wg notebooks leads kimwnasptd stefanofioravanzo elikatsis thesuperzapper could one of you create the pr to add the group please
1
170,644
6,468,228,823
IssuesEvent
2017-08-17 00:12:23
livro-aberto/BookCloud
https://api.github.com/repos/livro-aberto/BookCloud
closed
Edentação após Enter no admonition (e em outros ambientes também)
enhancement low priority user friendly
Ao pressionar Enter na linha abaixo o cursor vai para o início da linha debaixo, mas deveria ir direto para o terceiro espaço de duas linhas abaixo (edentação correta para se permanecer no ambiente). Acho que este era o problema que algumas pessoas reclamaram que o admonition não funcionava. `.. admonition:: Distratores posição atual do cursor após pressionar Enter` `.. admonition:: Distratores` `[linha vazia]` `[espaço, espaço, espaço] posição desejada do cursor após pressionar Enter` Como é um problema de simples solução que apareceu algumas vezes entre os elaboradores, vou colocar prioridade alta. Mas podemos rediscutir isso.
1.0
Edentação após Enter no admonition (e em outros ambientes também) - Ao pressionar Enter na linha abaixo o cursor vai para o início da linha debaixo, mas deveria ir direto para o terceiro espaço de duas linhas abaixo (edentação correta para se permanecer no ambiente). Acho que este era o problema que algumas pessoas reclamaram que o admonition não funcionava. `.. admonition:: Distratores posição atual do cursor após pressionar Enter` `.. admonition:: Distratores` `[linha vazia]` `[espaço, espaço, espaço] posição desejada do cursor após pressionar Enter` Como é um problema de simples solução que apareceu algumas vezes entre os elaboradores, vou colocar prioridade alta. Mas podemos rediscutir isso.
non_process
edentação após enter no admonition e em outros ambientes também ao pressionar enter na linha abaixo o cursor vai para o início da linha debaixo mas deveria ir direto para o terceiro espaço de duas linhas abaixo edentação correta para se permanecer no ambiente acho que este era o problema que algumas pessoas reclamaram que o admonition não funcionava admonition distratores posição atual do cursor após pressionar enter admonition distratores posição desejada do cursor após pressionar enter como é um problema de simples solução que apareceu algumas vezes entre os elaboradores vou colocar prioridade alta mas podemos rediscutir isso
0
17,025
22,392,788,776
IssuesEvent
2022-06-17 09:21:33
python/cpython
https://api.github.com/repos/python/cpython
closed
multiprocessing.Pool and ThreadPool leak resources after being deleted
type-bug docs stdlib 3.8 expert-multiprocessing
BPO | [34172](https://bugs.python.org/issue34172) --- | :--- Nosy | @pitrou, @vstinner, @benjaminp, @ned-deily, @zware, @mattip, @MojoVampire, @applio, @tzickel, @pablogsal, @Windsooon PRs | <li>python/cpython#8450</li><li>python/cpython#9676</li><li>python/cpython#9677</li><li>python/cpython#9686</li><li>python/cpython#10968</li><li>python/cpython#10969</li><li>python/cpython#10970</li><li>python/cpython#10971</li> <sup>*Note: these values reflect the state of the issue at the time it was migrated and might not reflect the current state.*</sup> <details><summary>Show more details</summary><p> GitHub fields: ```python assignee = None closed_at = None created_at = <Date 2018-07-20.16:45:29.720> labels = ['3.8', 'type-bug', 'library', 'docs'] title = 'multiprocessing.Pool and ThreadPool leak resources after being deleted' updated_at = <Date 2019-10-23.18:45:00.937> user = 'https://github.com/tzickel' ``` bugs.python.org fields: ```python activity = <Date 2019-10-23.18:45:00.937> actor = 'josh.r' assignee = 'docs@python' closed = False closed_date = None closer = None components = ['Documentation', 'Library (Lib)'] creation = <Date 2018-07-20.16:45:29.720> creator = 'tzickel' dependencies = [] files = [] hgrepos = [] issue_num = 34172 keywords = ['patch'] message_count = 52.0 messages = ['322028', '322076', '322090', '322129', '322147', '322150', '322172', '322190', '322194', '322230', '322231', '323103', '323104', '326904', '326909', '326913', '326914', '326915', '326948', '326964', '330864', '330865', '330866', '330868', '330869', '330891', '330954', '330955', '330960', '330962', '330988', '330996', '330997', '330999', '331009', '331026', '331087', '331190', '331198', '331199', '331200', '331210', '331216', '331218', '331221', '331629', '331703', '332294', '335357', '355229', '355247', '355248'] nosy_count = 13.0 nosy_names = ['pitrou', 'vstinner', 'benjamin.peterson', 'ned.deily', 'docs@python', 'zach.ware', 'mattip', 'josh.r', 'davin', 'tzickel', 'pablogsal', 'Windson Yang', 'Vy Nguyen'] pr_nums = ['8450', '9676', '9677', '9686', '10968', '10969', '10970', '10971'] priority = None resolution = None stage = None status = 'open' superseder = None type = 'behavior' url = 'https://bugs.python.org/issue34172' versions = ['Python 3.8'] ``` </p></details>
1.0
multiprocessing.Pool and ThreadPool leak resources after being deleted - BPO | [34172](https://bugs.python.org/issue34172) --- | :--- Nosy | @pitrou, @vstinner, @benjaminp, @ned-deily, @zware, @mattip, @MojoVampire, @applio, @tzickel, @pablogsal, @Windsooon PRs | <li>python/cpython#8450</li><li>python/cpython#9676</li><li>python/cpython#9677</li><li>python/cpython#9686</li><li>python/cpython#10968</li><li>python/cpython#10969</li><li>python/cpython#10970</li><li>python/cpython#10971</li> <sup>*Note: these values reflect the state of the issue at the time it was migrated and might not reflect the current state.*</sup> <details><summary>Show more details</summary><p> GitHub fields: ```python assignee = None closed_at = None created_at = <Date 2018-07-20.16:45:29.720> labels = ['3.8', 'type-bug', 'library', 'docs'] title = 'multiprocessing.Pool and ThreadPool leak resources after being deleted' updated_at = <Date 2019-10-23.18:45:00.937> user = 'https://github.com/tzickel' ``` bugs.python.org fields: ```python activity = <Date 2019-10-23.18:45:00.937> actor = 'josh.r' assignee = 'docs@python' closed = False closed_date = None closer = None components = ['Documentation', 'Library (Lib)'] creation = <Date 2018-07-20.16:45:29.720> creator = 'tzickel' dependencies = [] files = [] hgrepos = [] issue_num = 34172 keywords = ['patch'] message_count = 52.0 messages = ['322028', '322076', '322090', '322129', '322147', '322150', '322172', '322190', '322194', '322230', '322231', '323103', '323104', '326904', '326909', '326913', '326914', '326915', '326948', '326964', '330864', '330865', '330866', '330868', '330869', '330891', '330954', '330955', '330960', '330962', '330988', '330996', '330997', '330999', '331009', '331026', '331087', '331190', '331198', '331199', '331200', '331210', '331216', '331218', '331221', '331629', '331703', '332294', '335357', '355229', '355247', '355248'] nosy_count = 13.0 nosy_names = ['pitrou', 'vstinner', 'benjamin.peterson', 'ned.deily', 'docs@python', 'zach.ware', 'mattip', 'josh.r', 'davin', 'tzickel', 'pablogsal', 'Windson Yang', 'Vy Nguyen'] pr_nums = ['8450', '9676', '9677', '9686', '10968', '10969', '10970', '10971'] priority = None resolution = None stage = None status = 'open' superseder = None type = 'behavior' url = 'https://bugs.python.org/issue34172' versions = ['Python 3.8'] ``` </p></details>
process
multiprocessing pool and threadpool leak resources after being deleted bpo nosy pitrou vstinner benjaminp ned deily zware mattip mojovampire applio tzickel pablogsal windsooon prs python cpython python cpython python cpython python cpython python cpython python cpython python cpython python cpython note these values reflect the state of the issue at the time it was migrated and might not reflect the current state show more details github fields python assignee none closed at none created at labels title multiprocessing pool and threadpool leak resources after being deleted updated at user bugs python org fields python activity actor josh r assignee docs python closed false closed date none closer none components creation creator tzickel dependencies files hgrepos issue num keywords message count messages nosy count nosy names pr nums priority none resolution none stage none status open superseder none type behavior url versions
1
57,672
14,174,478,014
IssuesEvent
2020-11-12 19:58:09
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
build: Move the third party repositories to a c-deps/repositories.bzl rule
A-build-system C-cleanup
This is similar to how vendor is in there.
1.0
build: Move the third party repositories to a c-deps/repositories.bzl rule - This is similar to how vendor is in there.
non_process
build move the third party repositories to a c deps repositories bzl rule this is similar to how vendor is in there
0
176,313
13,635,310,865
IssuesEvent
2020-09-25 02:30:23
microsoft/AzureStorageExplorer
https://api.github.com/repos/microsoft/AzureStorageExplorer
closed
The localized string 'Search by prefix (case sensitive)' shows incompletely in Japanese
:gear: usability 🌐 localization 🧪 testing
**Storage Explorer Version:** 1.15.0-dev **Build**: 20200627.2 **Branch**: master **Language**: Japanese **Platform/OS:** Windows 10/ macOS Catalina **Architecture:** ia32/x64 **Regression From:** Not a regression **Steps to reproduce:** 1. Launch Storage Explorer. 2. Open 'Settings' -> Application (Regional Settings) -> Select '日本語' -> Restart Storage Explorer. 3. Expand one storage account -> Blob Containers. 4. Open one blob container -> Observe the localized string in **search by prefix** box. **Expect Experience:** The string shows completely. ![image](https://user-images.githubusercontent.com/41351993/85943608-b8b51f00-b963-11ea-9a4c-acff5db2866d.png) **Actual Experience:** The string shows incompletely. ![image](https://user-images.githubusercontent.com/41351993/85943603-b4890180-b963-11ea-8a88-a41504c562b5.png) **More Info:** This issue doesn't reproduce on Linux.
1.0
The localized string 'Search by prefix (case sensitive)' shows incompletely in Japanese - **Storage Explorer Version:** 1.15.0-dev **Build**: 20200627.2 **Branch**: master **Language**: Japanese **Platform/OS:** Windows 10/ macOS Catalina **Architecture:** ia32/x64 **Regression From:** Not a regression **Steps to reproduce:** 1. Launch Storage Explorer. 2. Open 'Settings' -> Application (Regional Settings) -> Select '日本語' -> Restart Storage Explorer. 3. Expand one storage account -> Blob Containers. 4. Open one blob container -> Observe the localized string in **search by prefix** box. **Expect Experience:** The string shows completely. ![image](https://user-images.githubusercontent.com/41351993/85943608-b8b51f00-b963-11ea-9a4c-acff5db2866d.png) **Actual Experience:** The string shows incompletely. ![image](https://user-images.githubusercontent.com/41351993/85943603-b4890180-b963-11ea-8a88-a41504c562b5.png) **More Info:** This issue doesn't reproduce on Linux.
non_process
the localized string search by prefix case sensitive shows incompletely in japanese storage explorer version dev build branch master language japanese platform os windows macos catalina architecture regression from not a regression steps to reproduce launch storage explorer open settings application regional settings select 日本語 restart storage explorer expand one storage account blob containers open one blob container observe the localized string in search by prefix box expect experience the string shows completely actual experience the string shows incompletely more info this issue doesn t reproduce on linux
0
182,502
14,138,370,934
IssuesEvent
2020-11-10 08:19:38
unicode-org/icu4x
https://api.github.com/repos/unicode-org/icu4x
closed
Exhaustive tests?
C-test-infra T-enhancement
In ICU, we have an exhaustive test mode for tests that are particularly slow. Examples of tests that might be exhaustive tests: 1. Test a feature over all locales\* 2. Test the bounds of something (e.g., longest supportive string) It would be nice if we had an exhaustive test mode in ICU4X, such that running `cargo test` remains fast and doesn't get stuck on slow tests. \* It's common in ICU to unconditionally test a random subset of locales, and test the full set in exhaustive mode.
1.0
Exhaustive tests? - In ICU, we have an exhaustive test mode for tests that are particularly slow. Examples of tests that might be exhaustive tests: 1. Test a feature over all locales\* 2. Test the bounds of something (e.g., longest supportive string) It would be nice if we had an exhaustive test mode in ICU4X, such that running `cargo test` remains fast and doesn't get stuck on slow tests. \* It's common in ICU to unconditionally test a random subset of locales, and test the full set in exhaustive mode.
non_process
exhaustive tests in icu we have an exhaustive test mode for tests that are particularly slow examples of tests that might be exhaustive tests test a feature over all locales test the bounds of something e g longest supportive string it would be nice if we had an exhaustive test mode in such that running cargo test remains fast and doesn t get stuck on slow tests it s common in icu to unconditionally test a random subset of locales and test the full set in exhaustive mode
0
19,852
26,254,420,643
IssuesEvent
2023-01-05 22:39:44
tradingstrategy-ai/frontend
https://api.github.com/repos/tradingstrategy-ai/frontend
closed
Add badge indicator to open/closed positions nav
enhancement process size: XS
Strategy pages that include the sub-nav should show a numeric badge indicator next to **Open positions** and **Closed positions** options. <img width="281" alt="image" src="https://user-images.githubusercontent.com/35901/207239539-8772b63e-3f77-44ba-a571-a670d3d92f63.png">
1.0
Add badge indicator to open/closed positions nav - Strategy pages that include the sub-nav should show a numeric badge indicator next to **Open positions** and **Closed positions** options. <img width="281" alt="image" src="https://user-images.githubusercontent.com/35901/207239539-8772b63e-3f77-44ba-a571-a670d3d92f63.png">
process
add badge indicator to open closed positions nav strategy pages that include the sub nav should show a numeric badge indicator next to open positions and closed positions options img width alt image src
1
231,516
17,692,100,830
IssuesEvent
2021-08-24 11:17:44
osquery/osquery
https://api.github.com/repos/osquery/osquery
opened
Update Linux installation default paths in code and in documentation
Linux documentation
Beyond a small issue with the symlinks in the Linux packages which I'm fixing in the `osquery-packaging`, we still need to update the documentation and default paths inside the code that still point under `/usr`.
1.0
Update Linux installation default paths in code and in documentation - Beyond a small issue with the symlinks in the Linux packages which I'm fixing in the `osquery-packaging`, we still need to update the documentation and default paths inside the code that still point under `/usr`.
non_process
update linux installation default paths in code and in documentation beyond a small issue with the symlinks in the linux packages which i m fixing in the osquery packaging we still need to update the documentation and default paths inside the code that still point under usr
0
374,608
11,092,992,869
IssuesEvent
2019-12-15 22:47:21
workcraft/workcraft
https://api.github.com/repos/workcraft/workcraft
closed
Separate Circuit preferences for active-high and active-low reset port
enhancement priority:low status:confirmed tag:model:circuit
Provide the means to configure default port names for active-high and active-low reset of Digital Circuit.
1.0
Separate Circuit preferences for active-high and active-low reset port - Provide the means to configure default port names for active-high and active-low reset of Digital Circuit.
non_process
separate circuit preferences for active high and active low reset port provide the means to configure default port names for active high and active low reset of digital circuit
0
376
2,823,449,635
IssuesEvent
2015-05-21 08:54:35
genomizer/genomizer-server
https://api.github.com/repos/genomizer/genomizer-server
closed
Cache timestamp breaks stuff
BL bug Data Storage Processing
?annotations=[expid] works fine but ?annotations=[expid]&_=12345 doesn't. Since the server doesn't cache, cache time stamps could just be thrown away.
1.0
Cache timestamp breaks stuff - ?annotations=[expid] works fine but ?annotations=[expid]&_=12345 doesn't. Since the server doesn't cache, cache time stamps could just be thrown away.
process
cache timestamp breaks stuff annotations works fine but annotations doesn t since the server doesn t cache cache time stamps could just be thrown away
1
3,878
17,190,195,281
IssuesEvent
2021-07-16 09:49:03
chocolatey-community/chocolatey-package-requests
https://api.github.com/repos/chocolatey-community/chocolatey-package-requests
closed
RFP - torifier
Status: Available For Maintainer(s) Status: Published
<!-- * Please ensure the package does not already exist in the Chocolatey Community Repository - https://chocolatey.org/packages - by using a relevant search. * Please ensure there is no existing open package request. * Please ensure the issue title starts with 'RFP - ' - for example 'RFP - adobe-reader' * Please also ensure the issue title matches the identifier you expect the package should be named. * Please ensure you have both the Software Project URL and the Software Download URL before continuing. NOTE: Keep in mind we have an etiquette regarding communication that we expect folks to observe when they are looking for support in the Chocolatey community - https://github.com/chocolatey/chocolatey-package-requests/blob/master/README.md#etiquette-regarding-communication PLEASE REMOVE ALL COMMENTS ONCE YOU HAVE READ THEM. --> ## Checklist - [x] The package I am requesting does not already exist on https://chocolatey.org/packages; - [x] There is no open issue for this package; - [x] The issue title starts with 'RFP - '; - [x] The download URL is public and not locked behind a paywall / login; ## Package Details Software project URL : https://www.torifier.com/ Direct download URL for the software / installer : https://cutt.ly/ebfiiPz Software summary / short description: tunnel software applications through Tor without the need to reconfigure them <!-- ## Package Expectations Here you can make suggestions on what you would expect the package to do outside of 'installing' - eg. adding icons to the desktop -->
True
RFP - torifier - <!-- * Please ensure the package does not already exist in the Chocolatey Community Repository - https://chocolatey.org/packages - by using a relevant search. * Please ensure there is no existing open package request. * Please ensure the issue title starts with 'RFP - ' - for example 'RFP - adobe-reader' * Please also ensure the issue title matches the identifier you expect the package should be named. * Please ensure you have both the Software Project URL and the Software Download URL before continuing. NOTE: Keep in mind we have an etiquette regarding communication that we expect folks to observe when they are looking for support in the Chocolatey community - https://github.com/chocolatey/chocolatey-package-requests/blob/master/README.md#etiquette-regarding-communication PLEASE REMOVE ALL COMMENTS ONCE YOU HAVE READ THEM. --> ## Checklist - [x] The package I am requesting does not already exist on https://chocolatey.org/packages; - [x] There is no open issue for this package; - [x] The issue title starts with 'RFP - '; - [x] The download URL is public and not locked behind a paywall / login; ## Package Details Software project URL : https://www.torifier.com/ Direct download URL for the software / installer : https://cutt.ly/ebfiiPz Software summary / short description: tunnel software applications through Tor without the need to reconfigure them <!-- ## Package Expectations Here you can make suggestions on what you would expect the package to do outside of 'installing' - eg. adding icons to the desktop -->
non_process
rfp torifier please ensure the package does not already exist in the chocolatey community repository by using a relevant search please ensure there is no existing open package request please ensure the issue title starts with rfp for example rfp adobe reader please also ensure the issue title matches the identifier you expect the package should be named please ensure you have both the software project url and the software download url before continuing note keep in mind we have an etiquette regarding communication that we expect folks to observe when they are looking for support in the chocolatey community please remove all comments once you have read them checklist the package i am requesting does not already exist on there is no open issue for this package the issue title starts with rfp the download url is public and not locked behind a paywall login package details software project url direct download url for the software installer software summary short description tunnel software applications through tor without the need to reconfigure them package expectations here you can make suggestions on what you would expect the package to do outside of installing eg adding icons to the desktop
0
119,554
25,536,611,720
IssuesEvent
2022-11-29 12:28:47
Regalis11/Barotrauma
https://api.github.com/repos/Regalis11/Barotrauma
closed
Iron helmet can be found in stores
Bug Code Design Unstable
### Disclaimers - [X] I have searched the issue tracker to check if the issue has already been reported. - [ ] My issue happened while using mods. ### What happened? ![image](https://user-images.githubusercontent.com/104232152/200136941-2c6603f8-1239-410a-9c4c-d04336f4386a.png) Simply, the makeshift armor isnt in stores anymore, the iron helmet still is. Happened on a residential Outpost, if that helps. ### Reproduction steps 1.Check stores for iron Helmets ### Bug prevalence Happens every time I play ### Version Faction test branch ### - _No response_ ### Which operating system did you encounter this bug on? Windows ### Relevant error messages and crash reports _No response_
1.0
Iron helmet can be found in stores - ### Disclaimers - [X] I have searched the issue tracker to check if the issue has already been reported. - [ ] My issue happened while using mods. ### What happened? ![image](https://user-images.githubusercontent.com/104232152/200136941-2c6603f8-1239-410a-9c4c-d04336f4386a.png) Simply, the makeshift armor isnt in stores anymore, the iron helmet still is. Happened on a residential Outpost, if that helps. ### Reproduction steps 1.Check stores for iron Helmets ### Bug prevalence Happens every time I play ### Version Faction test branch ### - _No response_ ### Which operating system did you encounter this bug on? Windows ### Relevant error messages and crash reports _No response_
non_process
iron helmet can be found in stores disclaimers i have searched the issue tracker to check if the issue has already been reported my issue happened while using mods what happened simply the makeshift armor isnt in stores anymore the iron helmet still is happened on a residential outpost if that helps reproduction steps check stores for iron helmets bug prevalence happens every time i play version faction test branch no response which operating system did you encounter this bug on windows relevant error messages and crash reports no response
0
25,813
11,220,137,236
IssuesEvent
2020-01-07 15:15:20
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
incorrect/outdated picture?
Pri2 assigned-to-author doc-enhancement security/subsvc sql-database/svc triaged
in the section "To set the Azure AD administrator: 1. In Azure portal, on the SQL server page, select Active Directory admin. Next select Set admin." You are showing the picture of the SQL databases &gt; yourDatabase blade But when going to my Azure account the blade 'SQL Databases'&gt;[mydatabase], I don't see any of the settings options, like 'manage backups' or 'active directory admin'. I only see that on the 'SQL servers' blade --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 91872d18-0fbe-7fca-8503-58fb1df06ad1 * Version Independent ID: d53dfe9e-9351-20d3-cbd4-c14617a40737 * Content: [Secure a single or pooled database - Azure SQL Database](https://docs.microsoft.com/en-us/azure/sql-database/sql-database-security-tutorial#feedback) * Content Source: [articles/sql-database/sql-database-security-tutorial.md](https://github.com/Microsoft/azure-docs/blob/master/articles/sql-database/sql-database-security-tutorial.md) * Service: **sql-database** * Sub-service: **security** * GitHub Login: @VanMSFT * Microsoft Alias: **vanto**
True
incorrect/outdated picture? - in the section "To set the Azure AD administrator: 1. In Azure portal, on the SQL server page, select Active Directory admin. Next select Set admin." You are showing the picture of the SQL databases &gt; yourDatabase blade But when going to my Azure account the blade 'SQL Databases'&gt;[mydatabase], I don't see any of the settings options, like 'manage backups' or 'active directory admin'. I only see that on the 'SQL servers' blade --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 91872d18-0fbe-7fca-8503-58fb1df06ad1 * Version Independent ID: d53dfe9e-9351-20d3-cbd4-c14617a40737 * Content: [Secure a single or pooled database - Azure SQL Database](https://docs.microsoft.com/en-us/azure/sql-database/sql-database-security-tutorial#feedback) * Content Source: [articles/sql-database/sql-database-security-tutorial.md](https://github.com/Microsoft/azure-docs/blob/master/articles/sql-database/sql-database-security-tutorial.md) * Service: **sql-database** * Sub-service: **security** * GitHub Login: @VanMSFT * Microsoft Alias: **vanto**
non_process
incorrect outdated picture in the section to set the azure ad administrator in azure portal on the sql server page select active directory admin next select set admin you are showing the picture of the sql databases gt yourdatabase blade but when going to my azure account the blade sql databases gt i don t see any of the settings options like manage backups or active directory admin i only see that on the sql servers blade document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service sql database sub service security github login vanmsft microsoft alias vanto
0
21,579
29,935,429,397
IssuesEvent
2023-06-22 12:27:35
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
opened
Status of Bazel 7.0.0-pre.20230615.1
P1 type: process release team-OSS
- Expected release date: 2023-06-22 Task list: - [x] Pick release baseline: [4e476b87](https://github.com/bazelbuild/bazel/commit/4e476b87d930a5b8ad0de09ecb75e5d7c45a47c6) - [ ] Create release candidate: https://releases.bazel.build/7.0.0/rolling/7.0.0-pre.20230615.1rc1/index.html - [ ] Post-submit: https://buildkite.com/bazel/bazel-bazel/builds?branch=release-7.0.0-pre.20230615.1rc1 - [ ] Push the release: https://releases.bazel.build/7.0.0/rolling/7.0.0-pre.20230615.1/index.html - [ ] Update the [release page](https://github.com/bazelbuild/bazel/releases/)
1.0
Status of Bazel 7.0.0-pre.20230615.1 - - Expected release date: 2023-06-22 Task list: - [x] Pick release baseline: [4e476b87](https://github.com/bazelbuild/bazel/commit/4e476b87d930a5b8ad0de09ecb75e5d7c45a47c6) - [ ] Create release candidate: https://releases.bazel.build/7.0.0/rolling/7.0.0-pre.20230615.1rc1/index.html - [ ] Post-submit: https://buildkite.com/bazel/bazel-bazel/builds?branch=release-7.0.0-pre.20230615.1rc1 - [ ] Push the release: https://releases.bazel.build/7.0.0/rolling/7.0.0-pre.20230615.1/index.html - [ ] Update the [release page](https://github.com/bazelbuild/bazel/releases/)
process
status of bazel pre expected release date task list pick release baseline create release candidate post submit push the release update the
1
8,658
11,797,947,717
IssuesEvent
2020-03-18 13:37:41
MHRA/products
https://api.github.com/repos/MHRA/products
opened
Observability | Structured logs for doc index updater
EPIC - Auto Batch Process :oncoming_automobile:
## User want Part of #408 As a tech user I want set structured logs for doc index updater So I can have meaningful data to monitor/alert/ debug ### Technical acceptance criteria Should have structured logs with this suggested schema ```json { "timestamp": "2020-03-18T13:13:05.442Z", "service_name": "medicines/doc-index-updater", "correlation_id": "92bb6eb5-b8c8-4ad3-8e5a-65bb71e34729", "event": { "label": "DOCUMENT_MANAGER_CREATE", "payload": { // } }, "level": "INFO" } ``` for the following events/interactions: - [ ] Sentinel <--> Document Manager - [ ] Document Manager <--> Service bus - [ ] Service bus <--> Check in - [ ] Check in <--> Sentinel file system - [ ] Check in <--> blob storage - [ ] Sentinel <--> State manager - [ ] State manager <--> Redis - [ ] State manager <--> Service bus - [ ] Serve logs from warp **Size** **Value** **Effort** ### Exit Criteria met - [ ] Backlog - [ ] Discovery - [ ] DUXD - [ ] Development - [ ] Quality Assurance - [ ] Release and Validate
1.0
Observability | Structured logs for doc index updater - ## User want Part of #408 As a tech user I want set structured logs for doc index updater So I can have meaningful data to monitor/alert/ debug ### Technical acceptance criteria Should have structured logs with this suggested schema ```json { "timestamp": "2020-03-18T13:13:05.442Z", "service_name": "medicines/doc-index-updater", "correlation_id": "92bb6eb5-b8c8-4ad3-8e5a-65bb71e34729", "event": { "label": "DOCUMENT_MANAGER_CREATE", "payload": { // } }, "level": "INFO" } ``` for the following events/interactions: - [ ] Sentinel <--> Document Manager - [ ] Document Manager <--> Service bus - [ ] Service bus <--> Check in - [ ] Check in <--> Sentinel file system - [ ] Check in <--> blob storage - [ ] Sentinel <--> State manager - [ ] State manager <--> Redis - [ ] State manager <--> Service bus - [ ] Serve logs from warp **Size** **Value** **Effort** ### Exit Criteria met - [ ] Backlog - [ ] Discovery - [ ] DUXD - [ ] Development - [ ] Quality Assurance - [ ] Release and Validate
process
observability structured logs for doc index updater user want part of as a tech user i want set structured logs for doc index updater so i can have meaningful data to monitor alert debug technical acceptance criteria should have structured logs with this suggested schema json timestamp service name medicines doc index updater correlation id event label document manager create payload level info for the following events interactions sentinel document manager document manager service bus service bus check in check in sentinel file system check in blob storage sentinel state manager state manager redis state manager service bus serve logs from warp size value effort exit criteria met backlog discovery duxd development quality assurance release and validate
1
21,370
29,202,227,103
IssuesEvent
2023-05-21 00:36:36
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
[Hibrido / Zona Norte de São Paulo - SP] C#/.Net Developer (Híbrido em Zona norte de SP) na Coodesh
SALVADOR BACK-END PJ SQL REQUISITOS PROCESSOS GITHUB SEGURANÇA UMA C QUALIDADE DESENVOLVIMENTO WEB ALOCADO Stale
## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/cnet-developer-l-211038204?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>A <strong>Beltis Tecnologia </strong>está em busca de <strong><ins>C#/.Net Developer</ins></strong> para compor seu time!</p> <p>Buscamos um profissional para ser parte integrante do time de desenvolvimento, capaz de desenhar novas soluções e sustentar os sistemas existentes. Buscamos uma pessoa dinâmica com capacidade para se adaptar a diferentes cenários solicitados. Você irá atuar em uma empresa do segmento bancário/financeiro de grande porte.</p> <p><strong>Responsabilidades:</strong></p> <ul> <li>Atuar no planejamento de novos produtos/serviços, atuando no desenvolvimento das interfaces para os usuários da plataforma;</li> <li>Atuar no planejamento e desenvolvimento de novas aplicações e funcionalidades;</li> <li>Fazer parte da sustentação e melhoria de projetos existentes;</li> <li>Desenvolver soluções técnicas;</li> <li>Testar as soluções desenvolvidas.</li> </ul> <p><strong>Motivos para fazer parte:</strong></p> <ul> <li>Projeto motivador em grande empresa do setor financeiro/bancário;</li> <li>Empresa que investe constantemente em tecnologias e novas soluções para garantir a evolução contínua da qualidade e segurança de seus sistemas;</li> <li>Ambiente colaborativo com grandes desafios;</li> <li>Nosso cliente está entre uma das melhores empresas na área de soluções financeiras e tecnologia;</li> <li>Parceria com plataforma de cursos online para aprimorar seus conhecimentos e habilidades.</li> </ul> ## BELTIS TECNOLOGIA: <p>Com mais de 20 anos de atuação no segmento de TI, adotamos uma política direcionada às pessoas. Com conhecimento de mercado, processos e tecnologia, oferecemos alta capacidade em Outsourcing de Profissionais.</p> <p>Atuamos com diversos clientes a nível nacional, grandes players em seus segmentos que vão desde o financeiro, varejo, ensino e órgãos públicos.</p><a href='https://coodesh.com/empresas/beltis-tecnologia'>Veja mais no site</a> ## Habilidades: - C# - .NET Framework - Microsoft SQL Server ## Local: Zona Norte de São Paulo - SP ## Requisitos: - Proficiente em desenvolvimento web; - Experiência com .NET Framework e C#; - Experiência com SQL Server; - Devido a questões de segurança as atividades só podem ser in loco. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [C#/.Net Developer (Híbrido em Zona norte de SP) na BELTIS TECNOLOGIA](https://coodesh.com/vagas/cnet-developer-l-211038204?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Alocado #### Regime PJ #### Categoria Back-End
1.0
[Hibrido / Zona Norte de São Paulo - SP] C#/.Net Developer (Híbrido em Zona norte de SP) na Coodesh - ## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/cnet-developer-l-211038204?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>A <strong>Beltis Tecnologia </strong>está em busca de <strong><ins>C#/.Net Developer</ins></strong> para compor seu time!</p> <p>Buscamos um profissional para ser parte integrante do time de desenvolvimento, capaz de desenhar novas soluções e sustentar os sistemas existentes. Buscamos uma pessoa dinâmica com capacidade para se adaptar a diferentes cenários solicitados. Você irá atuar em uma empresa do segmento bancário/financeiro de grande porte.</p> <p><strong>Responsabilidades:</strong></p> <ul> <li>Atuar no planejamento de novos produtos/serviços, atuando no desenvolvimento das interfaces para os usuários da plataforma;</li> <li>Atuar no planejamento e desenvolvimento de novas aplicações e funcionalidades;</li> <li>Fazer parte da sustentação e melhoria de projetos existentes;</li> <li>Desenvolver soluções técnicas;</li> <li>Testar as soluções desenvolvidas.</li> </ul> <p><strong>Motivos para fazer parte:</strong></p> <ul> <li>Projeto motivador em grande empresa do setor financeiro/bancário;</li> <li>Empresa que investe constantemente em tecnologias e novas soluções para garantir a evolução contínua da qualidade e segurança de seus sistemas;</li> <li>Ambiente colaborativo com grandes desafios;</li> <li>Nosso cliente está entre uma das melhores empresas na área de soluções financeiras e tecnologia;</li> <li>Parceria com plataforma de cursos online para aprimorar seus conhecimentos e habilidades.</li> </ul> ## BELTIS TECNOLOGIA: <p>Com mais de 20 anos de atuação no segmento de TI, adotamos uma política direcionada às pessoas. Com conhecimento de mercado, processos e tecnologia, oferecemos alta capacidade em Outsourcing de Profissionais.</p> <p>Atuamos com diversos clientes a nível nacional, grandes players em seus segmentos que vão desde o financeiro, varejo, ensino e órgãos públicos.</p><a href='https://coodesh.com/empresas/beltis-tecnologia'>Veja mais no site</a> ## Habilidades: - C# - .NET Framework - Microsoft SQL Server ## Local: Zona Norte de São Paulo - SP ## Requisitos: - Proficiente em desenvolvimento web; - Experiência com .NET Framework e C#; - Experiência com SQL Server; - Devido a questões de segurança as atividades só podem ser in loco. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [C#/.Net Developer (Híbrido em Zona norte de SP) na BELTIS TECNOLOGIA](https://coodesh.com/vagas/cnet-developer-l-211038204?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Alocado #### Regime PJ #### Categoria Back-End
process
c net developer híbrido em zona norte de sp na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 a beltis tecnologia está em busca de c net developer para compor seu time buscamos um profissional para ser parte integrante do time de desenvolvimento capaz de desenhar novas soluções e sustentar os sistemas existentes buscamos uma pessoa dinâmica com capacidade para se adaptar a diferentes cenários solicitados você irá atuar em uma empresa do segmento bancário financeiro de grande porte responsabilidades atuar no planejamento de novos produtos serviços atuando no desenvolvimento das interfaces para os usuários da plataforma atuar no planejamento e desenvolvimento de novas aplicações e funcionalidades fazer parte da sustentação e melhoria de projetos existentes desenvolver soluções técnicas testar as soluções desenvolvidas motivos para fazer parte projeto motivador em grande empresa do setor financeiro bancário empresa que investe constantemente em tecnologias e novas soluções para garantir a evolução contínua da qualidade e segurança de seus sistemas ambiente colaborativo com grandes desafios nosso cliente está entre uma das melhores empresas na área de soluções financeiras e tecnologia parceria com plataforma de cursos online para aprimorar seus conhecimentos e habilidades beltis tecnologia com mais de anos de atuação no segmento de ti adotamos uma política direcionada às pessoas com conhecimento de mercado processos e tecnologia oferecemos alta capacidade em outsourcing de profissionais atuamos com diversos clientes a nível nacional grandes players em seus segmentos que vão desde o financeiro varejo ensino e órgãos públicos habilidades c net framework microsoft sql server local zona norte de são paulo sp requisitos proficiente em desenvolvimento web experiência com net framework e c experiência com sql server devido a questões de segurança as atividades só podem ser in loco como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação alocado regime pj categoria back end
1
14,655
17,777,283,025
IssuesEvent
2021-08-30 21:01:13
habitat-sh/habitat
https://api.github.com/repos/habitat-sh/habitat
closed
Represent PIDs with a non-zero type
Type:Stability Focus:Supervisor ProcessManagement Stale Focus:Supervisor Focus:Launcher
Process IDs in the Supervisor should never be 0, so we should reflect that at the type level, using `std::num::NonZeroU32`.
1.0
Represent PIDs with a non-zero type - Process IDs in the Supervisor should never be 0, so we should reflect that at the type level, using `std::num::NonZeroU32`.
process
represent pids with a non zero type process ids in the supervisor should never be so we should reflect that at the type level using std num
1
139,736
11,276,661,790
IssuesEvent
2020-01-14 23:58:01
bcgov/range-web
https://api.github.com/repos/bcgov/range-web
closed
Delete option can not be chosen after cancelling a previous delete
bug ready to test
Similar to described in #147 (at least I think so) Applies to: - [x] Deleting a Pasture - [x] Deleting a plant community - [ ] Deleting Criteria indicator species (both range readiness and stubble height) - [x] Deleting a Minister’s issue
1.0
Delete option can not be chosen after cancelling a previous delete - Similar to described in #147 (at least I think so) Applies to: - [x] Deleting a Pasture - [x] Deleting a plant community - [ ] Deleting Criteria indicator species (both range readiness and stubble height) - [x] Deleting a Minister’s issue
non_process
delete option can not be chosen after cancelling a previous delete similar to described in at least i think so applies to deleting a pasture deleting a plant community deleting criteria indicator species both range readiness and stubble height deleting a minister’s issue
0
18,275
24,353,049,378
IssuesEvent
2022-10-03 03:25:02
prometheus-community/windows_exporter
https://api.github.com/repos/prometheus-community/windows_exporter
closed
windows_exporter [Failed to get the memory and CPU resource usage of the process]
collector/process info needed
## Failed to get the memory and CPU resource usage of the process. ```powershell .\windows_exporter-0.19.0-amd64.exe --collector.process.whitelist="Tabby.+|Google Chrome.+" ``` ![image](https://user-images.githubusercontent.com/93895686/191908030-48b38a86-3f8c-45d2-996f-b09ddabc6b89.png) ![image](https://user-images.githubusercontent.com/93895686/191908066-e572fad3-3c48-4d7a-9e1b-7d974482d505.png)
1.0
windows_exporter [Failed to get the memory and CPU resource usage of the process] - ## Failed to get the memory and CPU resource usage of the process. ```powershell .\windows_exporter-0.19.0-amd64.exe --collector.process.whitelist="Tabby.+|Google Chrome.+" ``` ![image](https://user-images.githubusercontent.com/93895686/191908030-48b38a86-3f8c-45d2-996f-b09ddabc6b89.png) ![image](https://user-images.githubusercontent.com/93895686/191908066-e572fad3-3c48-4d7a-9e1b-7d974482d505.png)
process
windows exporter failed to get the memory and cpu resource usage of the process powershell windows exporter exe collector process whitelist tabby google chrome
1
8,623
11,777,111,230
IssuesEvent
2020-03-16 14:20:15
prisma/prisma2
https://api.github.com/repos/prisma/prisma2
closed
TypeError: Cannot read property 'map' of undefined
process/candidate
Error occurs when starting up prisma after having successfully applied a schema. No queries are made, the server does not even become available as this error occurs. I'm not entirely sure how to resolve this issue, especially with us not being able to share the full schema, but it would be good to know how to debug this, or to find out what kind of schema setups would cause the "Cannot read property 'map' of undefined" error as right now it could be an actual bug, or an error in our schema, anywhere in the 4000+ lines of the schema. **package.json** for prisma looks like this: ``` { "name": "typescript-graphql", "scripts": { "start": "node dist/server", "clean": "rm -rf dist", "build": "npm -s run clean && npm -s run generate && tsc", "generate": "npm -s run generate:prisma && npm -s run generate:nexus", "generate:prisma": "prisma2 generate", "generate:nexus": "ts-node --transpile-only src/schema", "postinstall": "npm -s run generate", "dev": "ts-node-dev --no-notify --respawn --transpileOnly src/server" }, "dependencies": { "@prisma/client": "latest", "graphql": "14.6.0", "graphql-yoga": "1.18.3", "nexus": "0.12.0-rc.9", "nexus-prisma": "0.7.0" }, "devDependencies": { "@prisma/sdk": "0.0.172", "@types/node": "12.12.25", "@types/ws": "6.0.4", "prisma2": "latest", "ts-node": "8.5.4", "ts-node-dev": "1.0.0-pre.44", "typescript": "3.7.5" }, "prettier": { "singleQuote": true, "semi": false, "trailingComma": "all" }, "engines": { "node": ">=10.0.0" } } ``` Our **schema** is over 4000 lines long and contains information about the business that I should not share. It does though have a lot of relations like this: ``` model User { id Int @id createdAt DateTime @default(now()) email String @unique name String? role Role @default(USER) posts Post[] user1 User[] @relation("UserOne") user2 User[] @relation("UserTwo") } model Post { id Int @id createdAt DateTime @default(now()) updatedAt DateTime @updatedAt author User title String published Boolean @default(false) } enum Role { USER ADMIN } ``` **Migrating the database** `prisma2 migrate up --experimental` Outputs the schema, time it took etc. Can see the schema has been migrated successfully, tables, relations, indexes exist Start Typescript prisma2 express server: `npm run dev` Ouput: ``` > typescript-graphql@ dev /app/prisma/typescript/graphql > ts-node-dev --no-notify --respawn --transpileOnly src/server Using ts-node version 8.5.4, typescript version 3.7.5 TypeError: Cannot read property 'map' of undefined at Object.definition (/app/prisma/typescript/graphql/node_modules/nexus-prisma/src/publisher.ts:126:12) at SchemaBuilder.walkInputType (/app/prisma/typescript/graphql/node_modules/nexus/src/builder.ts:1363:9) at SchemaBuilder.walkTypes (/app/prisma/typescript/graphql/node_modules/nexus/src/builder.ts:678:16) at SchemaBuilder.getFinalTypeMap (/app/prisma/typescript/graphql/node_modules/nexus/src/builder.ts:786:10) at makeSchemaInternal (/app/prisma/typescript/graphql/node_modules/nexus/src/builder.ts:1583:15) at Object.makeSchema (/app/prisma/typescript/graphql/node_modules/nexus/src/builder.ts:1628:49) at Object.<anonymous> (/app/prisma/typescript/graphql/src/schema.ts:6:26) at Module._compile (internal/modules/cjs/loader.js:778:30) at Module._compile (/app/prisma/typescript/graphql/node_modules/source-map-support/source-map-support.js:541:25) at Module.m._compile (/tmp/ts-node-dev-hook-6002760669962068.js:57:25) [ERROR] 08:19:49 TypeError: Cannot read property 'map' of undefined ``` Adding DEBUG=* as environment variable did not provide more output.
1.0
TypeError: Cannot read property 'map' of undefined - Error occurs when starting up prisma after having successfully applied a schema. No queries are made, the server does not even become available as this error occurs. I'm not entirely sure how to resolve this issue, especially with us not being able to share the full schema, but it would be good to know how to debug this, or to find out what kind of schema setups would cause the "Cannot read property 'map' of undefined" error as right now it could be an actual bug, or an error in our schema, anywhere in the 4000+ lines of the schema. **package.json** for prisma looks like this: ``` { "name": "typescript-graphql", "scripts": { "start": "node dist/server", "clean": "rm -rf dist", "build": "npm -s run clean && npm -s run generate && tsc", "generate": "npm -s run generate:prisma && npm -s run generate:nexus", "generate:prisma": "prisma2 generate", "generate:nexus": "ts-node --transpile-only src/schema", "postinstall": "npm -s run generate", "dev": "ts-node-dev --no-notify --respawn --transpileOnly src/server" }, "dependencies": { "@prisma/client": "latest", "graphql": "14.6.0", "graphql-yoga": "1.18.3", "nexus": "0.12.0-rc.9", "nexus-prisma": "0.7.0" }, "devDependencies": { "@prisma/sdk": "0.0.172", "@types/node": "12.12.25", "@types/ws": "6.0.4", "prisma2": "latest", "ts-node": "8.5.4", "ts-node-dev": "1.0.0-pre.44", "typescript": "3.7.5" }, "prettier": { "singleQuote": true, "semi": false, "trailingComma": "all" }, "engines": { "node": ">=10.0.0" } } ``` Our **schema** is over 4000 lines long and contains information about the business that I should not share. It does though have a lot of relations like this: ``` model User { id Int @id createdAt DateTime @default(now()) email String @unique name String? role Role @default(USER) posts Post[] user1 User[] @relation("UserOne") user2 User[] @relation("UserTwo") } model Post { id Int @id createdAt DateTime @default(now()) updatedAt DateTime @updatedAt author User title String published Boolean @default(false) } enum Role { USER ADMIN } ``` **Migrating the database** `prisma2 migrate up --experimental` Outputs the schema, time it took etc. Can see the schema has been migrated successfully, tables, relations, indexes exist Start Typescript prisma2 express server: `npm run dev` Ouput: ``` > typescript-graphql@ dev /app/prisma/typescript/graphql > ts-node-dev --no-notify --respawn --transpileOnly src/server Using ts-node version 8.5.4, typescript version 3.7.5 TypeError: Cannot read property 'map' of undefined at Object.definition (/app/prisma/typescript/graphql/node_modules/nexus-prisma/src/publisher.ts:126:12) at SchemaBuilder.walkInputType (/app/prisma/typescript/graphql/node_modules/nexus/src/builder.ts:1363:9) at SchemaBuilder.walkTypes (/app/prisma/typescript/graphql/node_modules/nexus/src/builder.ts:678:16) at SchemaBuilder.getFinalTypeMap (/app/prisma/typescript/graphql/node_modules/nexus/src/builder.ts:786:10) at makeSchemaInternal (/app/prisma/typescript/graphql/node_modules/nexus/src/builder.ts:1583:15) at Object.makeSchema (/app/prisma/typescript/graphql/node_modules/nexus/src/builder.ts:1628:49) at Object.<anonymous> (/app/prisma/typescript/graphql/src/schema.ts:6:26) at Module._compile (internal/modules/cjs/loader.js:778:30) at Module._compile (/app/prisma/typescript/graphql/node_modules/source-map-support/source-map-support.js:541:25) at Module.m._compile (/tmp/ts-node-dev-hook-6002760669962068.js:57:25) [ERROR] 08:19:49 TypeError: Cannot read property 'map' of undefined ``` Adding DEBUG=* as environment variable did not provide more output.
process
typeerror cannot read property map of undefined error occurs when starting up prisma after having successfully applied a schema no queries are made the server does not even become available as this error occurs i m not entirely sure how to resolve this issue especially with us not being able to share the full schema but it would be good to know how to debug this or to find out what kind of schema setups would cause the cannot read property map of undefined error as right now it could be an actual bug or an error in our schema anywhere in the lines of the schema package json for prisma looks like this name typescript graphql scripts start node dist server clean rm rf dist build npm s run clean npm s run generate tsc generate npm s run generate prisma npm s run generate nexus generate prisma generate generate nexus ts node transpile only src schema postinstall npm s run generate dev ts node dev no notify respawn transpileonly src server dependencies prisma client latest graphql graphql yoga nexus rc nexus prisma devdependencies prisma sdk types node types ws latest ts node ts node dev pre typescript prettier singlequote true semi false trailingcomma all engines node our schema is over lines long and contains information about the business that i should not share it does though have a lot of relations like this model user id int id createdat datetime default now email string unique name string role role default user posts post user relation userone user relation usertwo model post id int id createdat datetime default now updatedat datetime updatedat author user title string published boolean default false enum role user admin migrating the database migrate up experimental outputs the schema time it took etc can see the schema has been migrated successfully tables relations indexes exist start typescript express server npm run dev ouput typescript graphql dev app prisma typescript graphql ts node dev no notify respawn transpileonly src server using ts node version typescript version typeerror cannot read property map of undefined at object definition app prisma typescript graphql node modules nexus prisma src publisher ts at schemabuilder walkinputtype app prisma typescript graphql node modules nexus src builder ts at schemabuilder walktypes app prisma typescript graphql node modules nexus src builder ts at schemabuilder getfinaltypemap app prisma typescript graphql node modules nexus src builder ts at makeschemainternal app prisma typescript graphql node modules nexus src builder ts at object makeschema app prisma typescript graphql node modules nexus src builder ts at object app prisma typescript graphql src schema ts at module compile internal modules cjs loader js at module compile app prisma typescript graphql node modules source map support source map support js at module m compile tmp ts node dev hook js typeerror cannot read property map of undefined adding debug as environment variable did not provide more output
1
9,026
2,615,122,023
IssuesEvent
2015-03-01 05:49:09
chrsmith/google-api-java-client
https://api.github.com/repos/chrsmith/google-api-java-client
closed
NullPointerException when non-top-level element declares a new namespace and uses it
auto-migrated Priority-Medium Type-Defect
``` Version of google-api-java-client (e.g. 1.4.1-beta)? 1.4.1-beta Java environment (e.g. Java 6, Android 2.3, App Engine 1.4.3)? JDK 1.6.0 Describe the problem. NullPointerException when namespace is defined on a non-top-level element Google Document List Feed produces an interior element <app:edited xmlns:app='http://www.w3.org/2007/app'>2011-08-09T04:38:14.017Z</app:edited> The xmlns:app definition only occurs on this element, not the top-level element. When this is parsed: java.lang.NullPointerException at com.google.api.client.xml.Xml.getFieldName(Xml.java:513) at com.google.api.client.xml.Xml.parseElementInternal(Xml.java:317) at com.google.api.client.xml.Xml.parseElementInternal(Xml.java:454) at com.google.api.client.xml.Xml.parseElement(Xml.java:198) at com.google.api.client.http.xml.XmlHttpParser.parse(XmlHttpParser.java:72) at com.google.api.client.http.HttpResponse.parseAs(HttpResponse.java:298) at org.springhaven.testing.GdataTest1.main(GdataTest1.java:166) How would you expect it to be fixed? Use a namespace stack and duplicate the namespace-addition code in Xml.java lines 228-247 at the start of each element (around line 317) to populate the stack. ``` Original issue reported on code.google.com by `darkfoxp...@gmail.com` on 13 Aug 2011 at 2:31
1.0
NullPointerException when non-top-level element declares a new namespace and uses it - ``` Version of google-api-java-client (e.g. 1.4.1-beta)? 1.4.1-beta Java environment (e.g. Java 6, Android 2.3, App Engine 1.4.3)? JDK 1.6.0 Describe the problem. NullPointerException when namespace is defined on a non-top-level element Google Document List Feed produces an interior element <app:edited xmlns:app='http://www.w3.org/2007/app'>2011-08-09T04:38:14.017Z</app:edited> The xmlns:app definition only occurs on this element, not the top-level element. When this is parsed: java.lang.NullPointerException at com.google.api.client.xml.Xml.getFieldName(Xml.java:513) at com.google.api.client.xml.Xml.parseElementInternal(Xml.java:317) at com.google.api.client.xml.Xml.parseElementInternal(Xml.java:454) at com.google.api.client.xml.Xml.parseElement(Xml.java:198) at com.google.api.client.http.xml.XmlHttpParser.parse(XmlHttpParser.java:72) at com.google.api.client.http.HttpResponse.parseAs(HttpResponse.java:298) at org.springhaven.testing.GdataTest1.main(GdataTest1.java:166) How would you expect it to be fixed? Use a namespace stack and duplicate the namespace-addition code in Xml.java lines 228-247 at the start of each element (around line 317) to populate the stack. ``` Original issue reported on code.google.com by `darkfoxp...@gmail.com` on 13 Aug 2011 at 2:31
non_process
nullpointerexception when non top level element declares a new namespace and uses it version of google api java client e g beta beta java environment e g java android app engine jdk describe the problem nullpointerexception when namespace is defined on a non top level element google document list feed produces an interior element app edited xmlns app the xmlns app definition only occurs on this element not the top level element when this is parsed java lang nullpointerexception at com google api client xml xml getfieldname xml java at com google api client xml xml parseelementinternal xml java at com google api client xml xml parseelementinternal xml java at com google api client xml xml parseelement xml java at com google api client http xml xmlhttpparser parse xmlhttpparser java at com google api client http httpresponse parseas httpresponse java at org springhaven testing main java how would you expect it to be fixed use a namespace stack and duplicate the namespace addition code in xml java lines at the start of each element around line to populate the stack original issue reported on code google com by darkfoxp gmail com on aug at
0
4,583
7,422,877,427
IssuesEvent
2018-03-23 01:54:29
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
What is "Properties" parameter value??
assigned-to-author azure-stack doc-bug in-process triaged
Missing the value~! --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 7dc8ef75-1f1d-2d0c-12ad-20159bfd26b9 * Version Independent ID: 03ecc613-7ab0-b470-6995-acc8583bb683 * Content: [Add tenants for usage and billing to Azure Stack](https://docs.microsoft.com/en-us/azure/azure-stack/azure-stack-csp-howto-register-tenants) * Content Source: [articles/azure-stack/azure-stack-csp-howto-register-tenants.md](https://github.com/Microsoft/azure-docs/blob/master/articles/azure-stack/azure-stack-csp-howto-register-tenants.md) * Service: **azure-stack** * GitHub Login: @mattbriggs * Microsoft Alias: **mabrigg**
1.0
What is "Properties" parameter value?? - Missing the value~! --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 7dc8ef75-1f1d-2d0c-12ad-20159bfd26b9 * Version Independent ID: 03ecc613-7ab0-b470-6995-acc8583bb683 * Content: [Add tenants for usage and billing to Azure Stack](https://docs.microsoft.com/en-us/azure/azure-stack/azure-stack-csp-howto-register-tenants) * Content Source: [articles/azure-stack/azure-stack-csp-howto-register-tenants.md](https://github.com/Microsoft/azure-docs/blob/master/articles/azure-stack/azure-stack-csp-howto-register-tenants.md) * Service: **azure-stack** * GitHub Login: @mattbriggs * Microsoft Alias: **mabrigg**
process
what is properties parameter value missing the value document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service azure stack github login mattbriggs microsoft alias mabrigg
1
4,058
6,992,780,876
IssuesEvent
2017-12-15 08:41:20
w3c/html
https://api.github.com/repos/w3c/html
closed
Identify sections at risk for 5.2
at risk help wanted meta process
As preparation for the HTML 5.2 Candidate Recommendation, we should identify anything in the specification that does not have interoperable implementation. Roughly, this means at least two working implementations that are **not** marked for deprecation - and ideally, some sense that further implementations are likely. Please list candidate sections in this issue, or open an issue suggesting something be marked at risk and noting why...
1.0
Identify sections at risk for 5.2 - As preparation for the HTML 5.2 Candidate Recommendation, we should identify anything in the specification that does not have interoperable implementation. Roughly, this means at least two working implementations that are **not** marked for deprecation - and ideally, some sense that further implementations are likely. Please list candidate sections in this issue, or open an issue suggesting something be marked at risk and noting why...
process
identify sections at risk for as preparation for the html candidate recommendation we should identify anything in the specification that does not have interoperable implementation roughly this means at least two working implementations that are not marked for deprecation and ideally some sense that further implementations are likely please list candidate sections in this issue or open an issue suggesting something be marked at risk and noting why
1
1,094
26,813,492,975
IssuesEvent
2023-02-02 01:10:32
openstates/issues
https://api.github.com/repos/openstates/issues
closed
New KY Committee Scraper
good first issue component:people-data good first scraper
### We need a new committee scraper written in [spatula](https://jamesturk.github.io/spatula/) for Kentucky. It should scrape this [webpage for Senate committees](https://legislature.ky.gov/Committees/senate-standing-committee), this [webpage for House committees](https://legislature.ky.gov/Committees/house-standing-committee), this [webpage for Joint committees](https://legislature.ky.gov/Committees/interim-joint-committee), this [webpage for Statutory committees](https://legislature.ky.gov/Committees/statutory-committee), and this [webpage for Special committees](https://legislature.ky.gov/Committees/special-committee) to get: - **name** - “Small Business”, “Agriculture”, etc. - **chamber** - “upper”, “lower”, or “legislature” (when joint, special, statutory, etc.) - **classification** - ex: "committee", "subcommittee" - **parent** - *only if it is a subcommittee* - scraper should have a way to determine this - ex: parent="Natural Resources" when *sub*committee name="Forestry" - **sources** - each added using `add_source()` method on instance of `ScrapeCommittee` type object - ex: home page for list of committees, specific page for that committee, etc. - **members** - - each added using `add_member()`method on instance of `ScrapeCommittee` type object - *Attributes:* - **name**: “Jane Doe”, “John Smith”, etc. - **role** (where applicable): “Chair”, “Ranking Member”, etc. #### [Here is helpful documentation](https://docs.openstates.org/contributing/writing-a-committee-scraper/) for writing a committee scraper ### Useful scrapers for reference An [old non-spatula-based scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers/ky/committees.py) that may be helpful in understanding how we have captured the data in the past. A [spatula-based scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers_next/mo/committees.py) that also scrapes an HTML List Page (using `HtmlListPage` and `HtmlPage` spatula classes) for reference. - As in the other spatula scraper, you will need to set the session variable separately (rather than depend on the `__init__.py`). ### Other useful resources You can reference the [object models for committee scrapers](https://github.com/openstates/openstates-core/blob/main/openstates/models/committees.py) in the Open States Core repo. Further documentation on running spatula scrapers in the command line can be found [here](https://docs.openstates.org/contributing/scrapers/#running-spatula-scrapers).
1.0
New KY Committee Scraper - ### We need a new committee scraper written in [spatula](https://jamesturk.github.io/spatula/) for Kentucky. It should scrape this [webpage for Senate committees](https://legislature.ky.gov/Committees/senate-standing-committee), this [webpage for House committees](https://legislature.ky.gov/Committees/house-standing-committee), this [webpage for Joint committees](https://legislature.ky.gov/Committees/interim-joint-committee), this [webpage for Statutory committees](https://legislature.ky.gov/Committees/statutory-committee), and this [webpage for Special committees](https://legislature.ky.gov/Committees/special-committee) to get: - **name** - “Small Business”, “Agriculture”, etc. - **chamber** - “upper”, “lower”, or “legislature” (when joint, special, statutory, etc.) - **classification** - ex: "committee", "subcommittee" - **parent** - *only if it is a subcommittee* - scraper should have a way to determine this - ex: parent="Natural Resources" when *sub*committee name="Forestry" - **sources** - each added using `add_source()` method on instance of `ScrapeCommittee` type object - ex: home page for list of committees, specific page for that committee, etc. - **members** - - each added using `add_member()`method on instance of `ScrapeCommittee` type object - *Attributes:* - **name**: “Jane Doe”, “John Smith”, etc. - **role** (where applicable): “Chair”, “Ranking Member”, etc. #### [Here is helpful documentation](https://docs.openstates.org/contributing/writing-a-committee-scraper/) for writing a committee scraper ### Useful scrapers for reference An [old non-spatula-based scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers/ky/committees.py) that may be helpful in understanding how we have captured the data in the past. A [spatula-based scraper](https://github.com/openstates/openstates-scrapers/blob/main/scrapers_next/mo/committees.py) that also scrapes an HTML List Page (using `HtmlListPage` and `HtmlPage` spatula classes) for reference. - As in the other spatula scraper, you will need to set the session variable separately (rather than depend on the `__init__.py`). ### Other useful resources You can reference the [object models for committee scrapers](https://github.com/openstates/openstates-core/blob/main/openstates/models/committees.py) in the Open States Core repo. Further documentation on running spatula scrapers in the command line can be found [here](https://docs.openstates.org/contributing/scrapers/#running-spatula-scrapers).
non_process
new ky committee scraper we need a new committee scraper written in for kentucky it should scrape this this this this and this to get name “small business” “agriculture” etc chamber “upper” “lower” or “legislature” when joint special statutory etc classification ex committee subcommittee parent only if it is a subcommittee scraper should have a way to determine this ex parent natural resources when sub committee name forestry sources each added using add source method on instance of scrapecommittee type object ex home page for list of committees specific page for that committee etc members each added using add member method on instance of scrapecommittee type object attributes name “jane doe” “john smith” etc role where applicable “chair” “ranking member” etc for writing a committee scraper useful scrapers for reference an that may be helpful in understanding how we have captured the data in the past a that also scrapes an html list page using htmllistpage and htmlpage spatula classes for reference as in the other spatula scraper you will need to set the session variable separately rather than depend on the init py other useful resources you can reference the in the open states core repo further documentation on running spatula scrapers in the command line can be found
0
120,314
17,644,085,941
IssuesEvent
2021-08-20 01:39:18
DavidSpek/pipelines
https://api.github.com/repos/DavidSpek/pipelines
opened
CVE-2021-37679 (High) detected in tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl
security vulnerability
## CVE-2021-37679 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/ec/98/f968caf5f65759e78873b900cbf0ae20b1699fb11268ecc0f892186419a7/tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/ec/98/f968caf5f65759e78873b900cbf0ae20b1699fb11268ecc0f892186419a7/tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl</a></p> <p>Path to dependency file: pipelines/contrib/components/openvino/ovms-deployer/containers/requirements.txt</p> <p>Path to vulnerable library: pipelines/contrib/components/openvino/ovms-deployer/containers/requirements.txt,pipelines/samples/core/ai_platform/training</p> <p> Dependency Hierarchy: - :x: **tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> TensorFlow is an end-to-end open source platform for machine learning. In affected versions it is possible to nest a `tf.map_fn` within another `tf.map_fn` call. However, if the input tensor is a `RaggedTensor` and there is no function signature provided, code assumes the output is a fully specified tensor and fills output buffer with uninitialized contents from the heap. The `t` and `z` outputs should be identical, however this is not the case. The last row of `t` contains data from the heap which can be used to leak other memory information. The bug lies in the conversion from a `Variant` tensor to a `RaggedTensor`. The [implementation](https://github.com/tensorflow/tensorflow/blob/460e000de3a83278fb00b61a16d161b1964f15f4/tensorflow/core/kernels/ragged_tensor_from_variant_op.cc#L177-L190) does not check that all inner shapes match and this results in the additional dimensions. The same implementation can result in data loss, if input tensor is tweaked. We have patched the issue in GitHub commit 4e2565483d0ffcadc719bd44893fb7f609bb5f12. The fix will be included in TensorFlow 2.6.0. We will also cherrypick this commit on TensorFlow 2.5.1, TensorFlow 2.4.3, and TensorFlow 2.3.4, as these are also affected and still in supported range. <p>Publish Date: 2021-08-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37679>CVE-2021-37679</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-g8wg-cjwc-xhhp">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-g8wg-cjwc-xhhp</a></p> <p>Release Date: 2021-08-12</p> <p>Fix Resolution: tensorflow - 2.3.4, 2.4.3, 2.5.1, 2.6.0, tensorflow-cpu - 2.3.4, 2.4.3, 2.5.1, 2.6.0, tensorflow-gpu - 2.3.4, 2.4.3, 2.5.1, 2.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-37679 (High) detected in tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl - ## CVE-2021-37679 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl</b></p></summary> <p>TensorFlow is an open source machine learning framework for everyone.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/ec/98/f968caf5f65759e78873b900cbf0ae20b1699fb11268ecc0f892186419a7/tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/ec/98/f968caf5f65759e78873b900cbf0ae20b1699fb11268ecc0f892186419a7/tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl</a></p> <p>Path to dependency file: pipelines/contrib/components/openvino/ovms-deployer/containers/requirements.txt</p> <p>Path to vulnerable library: pipelines/contrib/components/openvino/ovms-deployer/containers/requirements.txt,pipelines/samples/core/ai_platform/training</p> <p> Dependency Hierarchy: - :x: **tensorflow-1.15.0-cp27-cp27mu-manylinux2010_x86_64.whl** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> TensorFlow is an end-to-end open source platform for machine learning. In affected versions it is possible to nest a `tf.map_fn` within another `tf.map_fn` call. However, if the input tensor is a `RaggedTensor` and there is no function signature provided, code assumes the output is a fully specified tensor and fills output buffer with uninitialized contents from the heap. The `t` and `z` outputs should be identical, however this is not the case. The last row of `t` contains data from the heap which can be used to leak other memory information. The bug lies in the conversion from a `Variant` tensor to a `RaggedTensor`. The [implementation](https://github.com/tensorflow/tensorflow/blob/460e000de3a83278fb00b61a16d161b1964f15f4/tensorflow/core/kernels/ragged_tensor_from_variant_op.cc#L177-L190) does not check that all inner shapes match and this results in the additional dimensions. The same implementation can result in data loss, if input tensor is tweaked. We have patched the issue in GitHub commit 4e2565483d0ffcadc719bd44893fb7f609bb5f12. The fix will be included in TensorFlow 2.6.0. We will also cherrypick this commit on TensorFlow 2.5.1, TensorFlow 2.4.3, and TensorFlow 2.3.4, as these are also affected and still in supported range. <p>Publish Date: 2021-08-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-37679>CVE-2021-37679</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-g8wg-cjwc-xhhp">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-g8wg-cjwc-xhhp</a></p> <p>Release Date: 2021-08-12</p> <p>Fix Resolution: tensorflow - 2.3.4, 2.4.3, 2.5.1, 2.6.0, tensorflow-cpu - 2.3.4, 2.4.3, 2.5.1, 2.6.0, tensorflow-gpu - 2.3.4, 2.4.3, 2.5.1, 2.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in tensorflow whl cve high severity vulnerability vulnerable library tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file pipelines contrib components openvino ovms deployer containers requirements txt path to vulnerable library pipelines contrib components openvino ovms deployer containers requirements txt pipelines samples core ai platform training dependency hierarchy x tensorflow whl vulnerable library found in base branch master vulnerability details tensorflow is an end to end open source platform for machine learning in affected versions it is possible to nest a tf map fn within another tf map fn call however if the input tensor is a raggedtensor and there is no function signature provided code assumes the output is a fully specified tensor and fills output buffer with uninitialized contents from the heap the t and z outputs should be identical however this is not the case the last row of t contains data from the heap which can be used to leak other memory information the bug lies in the conversion from a variant tensor to a raggedtensor the does not check that all inner shapes match and this results in the additional dimensions the same implementation can result in data loss if input tensor is tweaked we have patched the issue in github commit the fix will be included in tensorflow we will also cherrypick this commit on tensorflow tensorflow and tensorflow as these are also affected and still in supported range publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow tensorflow cpu tensorflow gpu step up your open source security game with whitesource
0
4,242
7,187,127,872
IssuesEvent
2018-02-02 03:06:28
Great-Hill-Corporation/quickBlocks
https://api.github.com/repos/Great-Hill-Corporation/quickBlocks
closed
trace optimizations
monitors-all status-inprocess type-enhancement
If a trace has more than 50? 100? traces, then store the traces on the hard drive. 99% of the time needed to process a block with a big trace is taken by the node returning the traces and the javascript code parsing the JSON. Add a field to the transaction structure called 'tracesOnDisk' which instructs the monitor to read the traces from disc and then only load them if we're doing --deep
1.0
trace optimizations - If a trace has more than 50? 100? traces, then store the traces on the hard drive. 99% of the time needed to process a block with a big trace is taken by the node returning the traces and the javascript code parsing the JSON. Add a field to the transaction structure called 'tracesOnDisk' which instructs the monitor to read the traces from disc and then only load them if we're doing --deep
process
trace optimizations if a trace has more than traces then store the traces on the hard drive of the time needed to process a block with a big trace is taken by the node returning the traces and the javascript code parsing the json add a field to the transaction structure called tracesondisk which instructs the monitor to read the traces from disc and then only load them if we re doing deep
1
21,436
29,477,625,184
IssuesEvent
2023-06-02 00:38:08
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
[Remoto] Test Analyst na Coodesh
SALVADOR DESENVOLVIMENTO DE SOFTWARE GIT DEVOPS AWS REQUISITOS SELENIUM CUCUMBER REMOTO PROCESSOS GITHUB KANBAN SEGURANÇA UMA C CLEAN JEST APIs BARREIRAS TESTES AUTOMATIZADOS METODOLOGIAS ÁGEIS NEGÓCIOS AUTOMAÇÃO DE TESTES Stale
## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/jobs/test-analyst-160136060?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>O Grupo Fácil busca Test Analyst para compor seu time!</p> <p>Buscando projetos inovadores? Então vem com o Grupo Fácil e encontre a oportunidade certa. O Grupo Fácil oferece soluções e serviços inovadores em tecnologia da informação, para empresas dos mais diversos segmentos. O foco é proporcionar maior competitividade, agilidade e sustentabilidade para os clientes diante de um mercado cada vez mais desafiador.</p> <p>Quem somos nós:</p> <p>O Grupo Fácil sabe que a saúde é o pilar principal da vida das pessoas, por isso quebramos barreiras para promovê-la de forma sustentável junto a soluções inovadores para planos de saúde. Para realizar nosso propósito, inovamos todos os dias unindo pessoas, tecnologia e dados para prover o ecossistema da saúde com soluções que geram eficiência e sustentabilidade em toda a cadeia. Aqui, você irá participar, de maneira direta ou indireta, da criação de novas soluções para os mais diversos serviços de saúde, porque essa é a missão. Pesquisamos, criamos e fazemos acontecer!</p> <p>Como é trabalhar no Grupo Fácil:</p> <ul> <li>Nosso time é colaborativo e fazemos juntos para chegar no nosso objetivo;</li> <li>Vivemos agilidade e estamos em constante evolução;</li> <li>Estamos sempre buscando novas soluções inovadoras para o mercado;</li> </ul> <p>Sua missão como Analista de testes, será:</p> <ul> <li>Elaboração e execução de testes de APIS;</li> <li>Desenvolvimento de testes automatizados;</li> <li>Contribuir nos testes em desenvolvimentos de novas funcionalidades e ajustes das já existentes;</li> <li>Participar da criação de novos produtos inovadores aplicando as melhores práticas;</li> <li>Contribuir de forma colaborativa com time ágil;</li> <li>Colaborar em design de novos produtos em definições de novos escopos.</li> </ul> <p></p> ## Grupo Fácil: <p>Ao longo de 27 anos de história, o Grupo Fácil se tornou referência nacional em sistemas, softwares e serviços para a gestão de negócios nas áreas financeira e de crédito, da saúde e no setor imobiliário.</p> <p>O Grupo Fácil é formado por um conjunto de empresas que se destacam pela solidez e ousadia em projetos que otimizam processos e oferecem mais segurança e rentabilidade para seus clientes.&nbsp;</p><a href='https://coodesh.com/companies/grupo-facil'>Veja mais no site</a> ## Habilidades: - Selenium - Cucumber - Automação de Testes - API ## Local: 100% Remoto ## Requisitos: - Experiência em elaboração e execução de testes de APIS; - Experiência em testes funcionais e não funcionais de APIS; - Experiência em desenvolvimento de testes automatizados; - Experiência em desenvolvimento de testes integrados; - Experiência em testes de microserviços; - Conhecimento em Jest; - Consumo e testes em C#, .Net Core; - Versionamento com Git; - Experiência em clean code; - Experiência com Cloud pública AWS; - Experiência em técnicas de Devops; - Vivência em metodologias ágeis/Kanban; - Forte visão sistêmica (ciclo de desenvolvimento de software). ## Benefícios: - Convênio com farmácia; - Participação nos lucros; - Vale refeição; - Vale transporte; - Parcerias e convênios; - Programas de saúde e bem-estar. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Test Analyst na Grupo Fácil](https://coodesh.com/jobs/test-analyst-160136060?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Remoto #### Categoria Testes/Q.A
1.0
[Remoto] Test Analyst na Coodesh - ## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/jobs/test-analyst-160136060?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>O Grupo Fácil busca Test Analyst para compor seu time!</p> <p>Buscando projetos inovadores? Então vem com o Grupo Fácil e encontre a oportunidade certa. O Grupo Fácil oferece soluções e serviços inovadores em tecnologia da informação, para empresas dos mais diversos segmentos. O foco é proporcionar maior competitividade, agilidade e sustentabilidade para os clientes diante de um mercado cada vez mais desafiador.</p> <p>Quem somos nós:</p> <p>O Grupo Fácil sabe que a saúde é o pilar principal da vida das pessoas, por isso quebramos barreiras para promovê-la de forma sustentável junto a soluções inovadores para planos de saúde. Para realizar nosso propósito, inovamos todos os dias unindo pessoas, tecnologia e dados para prover o ecossistema da saúde com soluções que geram eficiência e sustentabilidade em toda a cadeia. Aqui, você irá participar, de maneira direta ou indireta, da criação de novas soluções para os mais diversos serviços de saúde, porque essa é a missão. Pesquisamos, criamos e fazemos acontecer!</p> <p>Como é trabalhar no Grupo Fácil:</p> <ul> <li>Nosso time é colaborativo e fazemos juntos para chegar no nosso objetivo;</li> <li>Vivemos agilidade e estamos em constante evolução;</li> <li>Estamos sempre buscando novas soluções inovadoras para o mercado;</li> </ul> <p>Sua missão como Analista de testes, será:</p> <ul> <li>Elaboração e execução de testes de APIS;</li> <li>Desenvolvimento de testes automatizados;</li> <li>Contribuir nos testes em desenvolvimentos de novas funcionalidades e ajustes das já existentes;</li> <li>Participar da criação de novos produtos inovadores aplicando as melhores práticas;</li> <li>Contribuir de forma colaborativa com time ágil;</li> <li>Colaborar em design de novos produtos em definições de novos escopos.</li> </ul> <p></p> ## Grupo Fácil: <p>Ao longo de 27 anos de história, o Grupo Fácil se tornou referência nacional em sistemas, softwares e serviços para a gestão de negócios nas áreas financeira e de crédito, da saúde e no setor imobiliário.</p> <p>O Grupo Fácil é formado por um conjunto de empresas que se destacam pela solidez e ousadia em projetos que otimizam processos e oferecem mais segurança e rentabilidade para seus clientes.&nbsp;</p><a href='https://coodesh.com/companies/grupo-facil'>Veja mais no site</a> ## Habilidades: - Selenium - Cucumber - Automação de Testes - API ## Local: 100% Remoto ## Requisitos: - Experiência em elaboração e execução de testes de APIS; - Experiência em testes funcionais e não funcionais de APIS; - Experiência em desenvolvimento de testes automatizados; - Experiência em desenvolvimento de testes integrados; - Experiência em testes de microserviços; - Conhecimento em Jest; - Consumo e testes em C#, .Net Core; - Versionamento com Git; - Experiência em clean code; - Experiência com Cloud pública AWS; - Experiência em técnicas de Devops; - Vivência em metodologias ágeis/Kanban; - Forte visão sistêmica (ciclo de desenvolvimento de software). ## Benefícios: - Convênio com farmácia; - Participação nos lucros; - Vale refeição; - Vale transporte; - Parcerias e convênios; - Programas de saúde e bem-estar. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Test Analyst na Grupo Fácil](https://coodesh.com/jobs/test-analyst-160136060?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Remoto #### Categoria Testes/Q.A
process
test analyst na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 o grupo fácil busca test analyst para compor seu time buscando projetos inovadores então vem com o grupo fácil e encontre a oportunidade certa o grupo fácil oferece soluções e serviços inovadores em tecnologia da informação para empresas dos mais diversos segmentos o foco é proporcionar maior competitividade agilidade e sustentabilidade para os clientes diante de um mercado cada vez mais desafiador quem somos nós o grupo fácil sabe que a saúde é o pilar principal da vida das pessoas por isso quebramos barreiras para promovê la de forma sustentável junto a soluções inovadores para planos de saúde para realizar nosso propósito inovamos todos os dias unindo pessoas tecnologia e dados para prover o ecossistema da saúde com soluções que geram eficiência e sustentabilidade em toda a cadeia aqui você irá participar de maneira direta ou indireta da criação de novas soluções para os mais diversos serviços de saúde porque essa é a missão pesquisamos criamos e fazemos acontecer como é trabalhar no grupo fácil nosso time é colaborativo e fazemos juntos para chegar no nosso objetivo vivemos agilidade e estamos em constante evolução estamos sempre buscando novas soluções inovadoras para o mercado sua missão como analista de testes será elaboração e execução de testes de apis desenvolvimento de testes automatizados contribuir nos testes em desenvolvimentos de novas funcionalidades e ajustes das já existentes participar da criação de novos produtos inovadores aplicando as melhores práticas contribuir de forma colaborativa com time ágil colaborar em design de novos produtos em definições de novos escopos grupo fácil ao longo de anos de história o grupo fácil se tornou referência nacional em sistemas softwares e serviços para a gestão de negócios nas áreas financeira e de crédito da saúde e no setor imobiliário o grupo fácil é formado por um conjunto de empresas que se destacam pela solidez e ousadia em projetos que otimizam processos e oferecem mais segurança e rentabilidade para seus clientes nbsp habilidades selenium cucumber automação de testes api local remoto requisitos experiência em elaboração e execução de testes de apis experiência em testes funcionais e não funcionais de apis experiência em desenvolvimento de testes automatizados experiência em desenvolvimento de testes integrados experiência em testes de microserviços conhecimento em jest consumo e testes em c net core versionamento com git experiência em clean code experiência com cloud pública aws experiência em técnicas de devops vivência em metodologias ágeis kanban forte visão sistêmica ciclo de desenvolvimento de software benefícios convênio com farmácia participação nos lucros vale refeição vale transporte parcerias e convênios programas de saúde e bem estar como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação remoto categoria testes q a
1
9,442
12,426,384,159
IssuesEvent
2020-05-24 20:54:21
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
Download options of Charts as PNG or JPEG format
.Proposal Querying/Processor Reporting/Pulses Type:New Feature Visualization/Chart Settings
The generated charts in questions pane shall include export options in jpeg/png format (as in csv and xlsx). Export functionality for questions may include the following options: - Transparent background (for png) - Resolution options - Include a link showing the URL of question below the chart as reference - Export time, and the email of person who exports the chart - Night mode / normal mode
1.0
Download options of Charts as PNG or JPEG format - The generated charts in questions pane shall include export options in jpeg/png format (as in csv and xlsx). Export functionality for questions may include the following options: - Transparent background (for png) - Resolution options - Include a link showing the URL of question below the chart as reference - Export time, and the email of person who exports the chart - Night mode / normal mode
process
download options of charts as png or jpeg format the generated charts in questions pane shall include export options in jpeg png format as in csv and xlsx export functionality for questions may include the following options transparent background for png resolution options include a link showing the url of question below the chart as reference export time and the email of person who exports the chart night mode normal mode
1