Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
757
labels
stringlengths
4
664
body
stringlengths
3
261k
index
stringclasses
10 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
232k
binary_label
int64
0
1
19,878
4,454,497,887
IssuesEvent
2016-08-23 01:08:48
evhub/coconut
https://api.github.com/repos/evhub/coconut
closed
Improve iterator matching documentation
documentation
Because Python iterators aren't real lazy lists, you have to be very careful about accidentally consuming them when performing checks, which makes iterator matching rather confusing.
1.0
Improve iterator matching documentation - Because Python iterators aren't real lazy lists, you have to be very careful about accidentally consuming them when performing checks, which makes iterator matching rather confusing.
non_defect
improve iterator matching documentation because python iterators aren t real lazy lists you have to be very careful about accidentally consuming them when performing checks which makes iterator matching rather confusing
0
172,637
13,325,483,657
IssuesEvent
2020-08-27 10:02:00
gambitph/Stackable
https://api.github.com/repos/gambitph/Stackable
opened
Social Snap Plugin "Pin It" feature is adding some custom CSS code to the image of the Testimonial block making the image small
[block] testimonial bug
Social Snap Plugin "Pin It" feature adds a Pinterest button overall images, making it small. ![image](https://user-images.githubusercontent.com/51441886/91426622-56a76600-e88f-11ea-9072-a4357a788541.png) Here's the code to fix the issue: `.ss-on-media-image-wrap.ugb-img--shape { display: block !important; }`
1.0
Social Snap Plugin "Pin It" feature is adding some custom CSS code to the image of the Testimonial block making the image small - Social Snap Plugin "Pin It" feature adds a Pinterest button overall images, making it small. ![image](https://user-images.githubusercontent.com/51441886/91426622-56a76600-e88f-11ea-9072-a4357a788541.png) Here's the code to fix the issue: `.ss-on-media-image-wrap.ugb-img--shape { display: block !important; }`
non_defect
social snap plugin pin it feature is adding some custom css code to the image of the testimonial block making the image small social snap plugin pin it feature adds a pinterest button overall images making it small here s the code to fix the issue ss on media image wrap ugb img shape display block important
0
81,023
30,662,370,985
IssuesEvent
2023-07-25 15:47:22
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
Hazelcast documentation incorrectly states that "local" option is available for ItemListener [HZ-2777]
Type: Defect Team: Core Type: Cleanup Source: Community Module: IQueue Module: ISet Module: ReplicatedMap Module: IList Type: Documentation to-jira
<!-- Thanks for reporting your issue. Please share with us the following information, to help us resolve your issue quickly and efficiently. --> **Describe the bug** The documentation states: `item-listeners: Lets you add listeners (listener classes) for the list items. ... You can set **local** to true if you want to listen to the items on the local member.` **Expected behavior** The **local** option is not available for IQueue, ISet, IList or ReplicatedMap ItemListener. **To Reproduce** instance.getQueue("someQueue") .addItemListener(new MyItemListener<>(), true); No option to specify if the listener should only run on local member that hosts the queue.
1.0
Hazelcast documentation incorrectly states that "local" option is available for ItemListener [HZ-2777] - <!-- Thanks for reporting your issue. Please share with us the following information, to help us resolve your issue quickly and efficiently. --> **Describe the bug** The documentation states: `item-listeners: Lets you add listeners (listener classes) for the list items. ... You can set **local** to true if you want to listen to the items on the local member.` **Expected behavior** The **local** option is not available for IQueue, ISet, IList or ReplicatedMap ItemListener. **To Reproduce** instance.getQueue("someQueue") .addItemListener(new MyItemListener<>(), true); No option to specify if the listener should only run on local member that hosts the queue.
defect
hazelcast documentation incorrectly states that local option is available for itemlistener thanks for reporting your issue please share with us the following information to help us resolve your issue quickly and efficiently describe the bug the documentation states item listeners lets you add listeners listener classes for the list items you can set local to true if you want to listen to the items on the local member expected behavior the local option is not available for iqueue iset ilist or replicatedmap itemlistener to reproduce instance getqueue somequeue additemlistener new myitemlistener true no option to specify if the listener should only run on local member that hosts the queue
1
35,338
7,702,684,008
IssuesEvent
2018-05-21 04:13:50
gperftools/gperftools
https://api.github.com/repos/gperftools/gperftools
closed
Option to disable writing tcmalloc larg allocation messages to stderr
Priority-Medium Status-WontFix Type-Defect
Originally reported on Google Code with ID 357 ``` tcmalloc will print messages such as the following to stderr: tcmalloc: large alloc 1176436736 bytes == 0x3177a000 @ 0x7f909f9fce6c 0x401ef9 0x40b931 0x44f1da 0x44be5f 0x4502c3 0x7f909ed9bc4d 0x401db9 Looking at the code, the output to stderr looks built in. Is there a mechanism for turning this off, as not to contaminate stderr unless there is actually an error? ``` Reported by `nilshomer` on 2011-07-29 19:43:15
1.0
Option to disable writing tcmalloc larg allocation messages to stderr - Originally reported on Google Code with ID 357 ``` tcmalloc will print messages such as the following to stderr: tcmalloc: large alloc 1176436736 bytes == 0x3177a000 @ 0x7f909f9fce6c 0x401ef9 0x40b931 0x44f1da 0x44be5f 0x4502c3 0x7f909ed9bc4d 0x401db9 Looking at the code, the output to stderr looks built in. Is there a mechanism for turning this off, as not to contaminate stderr unless there is actually an error? ``` Reported by `nilshomer` on 2011-07-29 19:43:15
defect
option to disable writing tcmalloc larg allocation messages to stderr originally reported on google code with id tcmalloc will print messages such as the following to stderr tcmalloc large alloc bytes looking at the code the output to stderr looks built in is there a mechanism for turning this off as not to contaminate stderr unless there is actually an error reported by nilshomer on
1
49,010
13,185,192,482
IssuesEvent
2020-08-12 20:54:28
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
check for moved/copied tools (Trac #598)
Incomplete Migration Migrated from Trac cmake defect
<details> <summary><em>Migrated from https://code.icecube.wisc.edu/ticket/598 , reported by troy and owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2012-06-17T19:55:06", "description": "not clear when to check, perhaps in the 'port' command and when running cmake.\n\n{{{\n\n14:15 <molivo> it says \"qmake found at bla bla bla\" then two warnings QT_QMAKE_EXECUTABLE reported QT_INSTALL_LIBS ...\n14:15 <molivo> wait a sec i am a bit messed up :-)\n14:20 <molivo> -- qmake found at /rest/tools/qt-4.6.0/bin/qmake\n14:20 <molivo> Warning: QT_QMAKE_EXECUTABLE reported QT_INSTALL_LIBS as /rest/icecube/tools/qt-4.6.0/lib\n14:20 <molivo> Warning: /rest/icecube/tools/qt-4.6.0/lib does NOT exist, Qt must NOT be installed correctly.\n14:20 <molivo> this is what I get\n14:21 <straszhm> hm, and that directory does exist... you said it contains libQtOpenGL.so\n14:21 <straszhm> (?)\n14:21 <molivo> yes it does\n14:22 <straszhm> random try: rm your CMakeCache.txt and run cmake again\n14:22 <molivo> ok\n14:22 <straszhm> hrm\n14:22 <straszhm> wait,\n14:22 <straszhm> -- qmake found at /rest/tools/qt-4.6.0/bin/qmake\n14:22 <molivo> yes\n14:22 <straszhm> reported QT_INSTALL_LIBS as /rest/icecube/tools/qt-4.6.0/lib\n14:22 <straszhm> those are two different directories\n14:23 <molivo> yes...\n14:23 <straszhm> did you move the toolset after you built it?\n14:23 <molivo> well, I did not. I copied from jens' pc\n14:23 <molivo> maybe he did\n}}}", "reporter": "troy", "cc": "", "resolution": "wont or cant fix", "_ts": "1339962906000000", "component": "cmake", "summary": "check for moved/copied tools", "priority": "normal", "keywords": "", "time": "2010-02-23T19:26:18", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
1.0
check for moved/copied tools (Trac #598) - <details> <summary><em>Migrated from https://code.icecube.wisc.edu/ticket/598 , reported by troy and owned by nega</em></summary> <p> ```json { "status": "closed", "changetime": "2012-06-17T19:55:06", "description": "not clear when to check, perhaps in the 'port' command and when running cmake.\n\n{{{\n\n14:15 <molivo> it says \"qmake found at bla bla bla\" then two warnings QT_QMAKE_EXECUTABLE reported QT_INSTALL_LIBS ...\n14:15 <molivo> wait a sec i am a bit messed up :-)\n14:20 <molivo> -- qmake found at /rest/tools/qt-4.6.0/bin/qmake\n14:20 <molivo> Warning: QT_QMAKE_EXECUTABLE reported QT_INSTALL_LIBS as /rest/icecube/tools/qt-4.6.0/lib\n14:20 <molivo> Warning: /rest/icecube/tools/qt-4.6.0/lib does NOT exist, Qt must NOT be installed correctly.\n14:20 <molivo> this is what I get\n14:21 <straszhm> hm, and that directory does exist... you said it contains libQtOpenGL.so\n14:21 <straszhm> (?)\n14:21 <molivo> yes it does\n14:22 <straszhm> random try: rm your CMakeCache.txt and run cmake again\n14:22 <molivo> ok\n14:22 <straszhm> hrm\n14:22 <straszhm> wait,\n14:22 <straszhm> -- qmake found at /rest/tools/qt-4.6.0/bin/qmake\n14:22 <molivo> yes\n14:22 <straszhm> reported QT_INSTALL_LIBS as /rest/icecube/tools/qt-4.6.0/lib\n14:22 <straszhm> those are two different directories\n14:23 <molivo> yes...\n14:23 <straszhm> did you move the toolset after you built it?\n14:23 <molivo> well, I did not. I copied from jens' pc\n14:23 <molivo> maybe he did\n}}}", "reporter": "troy", "cc": "", "resolution": "wont or cant fix", "_ts": "1339962906000000", "component": "cmake", "summary": "check for moved/copied tools", "priority": "normal", "keywords": "", "time": "2010-02-23T19:26:18", "milestone": "", "owner": "nega", "type": "defect" } ``` </p> </details>
defect
check for moved copied tools trac migrated from reported by troy and owned by nega json status closed changetime description not clear when to check perhaps in the port command and when running cmake n n n it says qmake found at bla bla bla then two warnings qt qmake executable reported qt install libs wait a sec i am a bit messed up qmake found at rest tools qt bin qmake warning qt qmake executable reported qt install libs as rest icecube tools qt lib warning rest icecube tools qt lib does not exist qt must not be installed correctly this is what i get hm and that directory does exist you said it contains libqtopengl so yes it does random try rm your cmakecache txt and run cmake again ok hrm wait qmake found at rest tools qt bin qmake yes reported qt install libs as rest icecube tools qt lib those are two different directories yes did you move the toolset after you built it well i did not i copied from jens pc maybe he did n reporter troy cc resolution wont or cant fix ts component cmake summary check for moved copied tools priority normal keywords time milestone owner nega type defect
1
39,273
9,368,595,978
IssuesEvent
2019-04-03 09:03:52
Automattic/wp-calypso
https://api.github.com/repos/Automattic/wp-calypso
closed
SDK: Build source outside Calypso repo
[Project] SDK [Type] Defect
The SDK should be able to build from source outside of Calypso. Currently, it cannot. #### Steps to reproduce Try to build some source that isn't inside of Calypso: ```sh npm run sdk generic /some/path/to/some/file.js output/file.js # Module not found: Error: Can't resolve '@babel/runtime/helpers/typeof' in … ``` It fails cc: @Automattic/calypso-sdk
1.0
SDK: Build source outside Calypso repo - The SDK should be able to build from source outside of Calypso. Currently, it cannot. #### Steps to reproduce Try to build some source that isn't inside of Calypso: ```sh npm run sdk generic /some/path/to/some/file.js output/file.js # Module not found: Error: Can't resolve '@babel/runtime/helpers/typeof' in … ``` It fails cc: @Automattic/calypso-sdk
defect
sdk build source outside calypso repo the sdk should be able to build from source outside of calypso currently it cannot steps to reproduce try to build some source that isn t inside of calypso sh npm run sdk generic some path to some file js output file js module not found error can t resolve babel runtime helpers typeof in … it fails cc automattic calypso sdk
1
431,282
12,476,697,067
IssuesEvent
2020-05-29 13:54:40
medic/cht-core
https://api.github.com/repos/medic/cht-core
closed
Support resolving tasks concerning unknown contacts
Priority: 1 - High Type: Bug
**Describe the bug** We can no longer rely on `action.modifyContent` to inject contact information into reports generated by task actions. This causes tasks that are about unknown contacts to generate reports that are unattached (are missing their subject). This workflow is used where supervisors with replication depth receive `needs_signoff` reports about patients they can't see. **To Reproduce** 1. Create a supervisor with replication_depth low enough that they don't see patients. 2. Create a chw under this supervisor. 3. Configure a form with needs_signoff. 4. Configure a task for the supervisor triggered by the `needs_signoff` form. Configure this to be resolved if the task action is executed. 5. Submit the `needs_signoff` form as the CHW. 6. Sync as the supervisor and check if you have generated the task. 7. Complete the task. 8. Your task does not resolve (it's still in the list) and inspecting the resulting report, it's missing all patient information. **Expected behavior** The resulting report should have the same patient information as the generating report. **Environment** - Instance: local - App: webapp - Version: 3.9 **Additional context** When written to disk, tasks are minified and they only retain a "forId" property representing `action.contact._id`, while `action.contact` is deleted. https://github.com/medic/cht-core/blob/master/shared-libs/rules-engine/src/transform-task-emission-to-doc.js#L77 When rehydrating the task, we rely on the `forId` contact to exist in order to repopulate `action.contact`: https://github.com/medic/cht-core/blob/master/webapp/src/js/controllers/tasks-content.js#L109 We should either: a) keep `action.contact` when the contact is unknown - this allows more information besides a uuid to be injected into the resulting report b) use a stub when rehydrating and contact is unknown - this means that the resulting report will only have the uuid available and no other information.
1.0
Support resolving tasks concerning unknown contacts - **Describe the bug** We can no longer rely on `action.modifyContent` to inject contact information into reports generated by task actions. This causes tasks that are about unknown contacts to generate reports that are unattached (are missing their subject). This workflow is used where supervisors with replication depth receive `needs_signoff` reports about patients they can't see. **To Reproduce** 1. Create a supervisor with replication_depth low enough that they don't see patients. 2. Create a chw under this supervisor. 3. Configure a form with needs_signoff. 4. Configure a task for the supervisor triggered by the `needs_signoff` form. Configure this to be resolved if the task action is executed. 5. Submit the `needs_signoff` form as the CHW. 6. Sync as the supervisor and check if you have generated the task. 7. Complete the task. 8. Your task does not resolve (it's still in the list) and inspecting the resulting report, it's missing all patient information. **Expected behavior** The resulting report should have the same patient information as the generating report. **Environment** - Instance: local - App: webapp - Version: 3.9 **Additional context** When written to disk, tasks are minified and they only retain a "forId" property representing `action.contact._id`, while `action.contact` is deleted. https://github.com/medic/cht-core/blob/master/shared-libs/rules-engine/src/transform-task-emission-to-doc.js#L77 When rehydrating the task, we rely on the `forId` contact to exist in order to repopulate `action.contact`: https://github.com/medic/cht-core/blob/master/webapp/src/js/controllers/tasks-content.js#L109 We should either: a) keep `action.contact` when the contact is unknown - this allows more information besides a uuid to be injected into the resulting report b) use a stub when rehydrating and contact is unknown - this means that the resulting report will only have the uuid available and no other information.
non_defect
support resolving tasks concerning unknown contacts describe the bug we can no longer rely on action modifycontent to inject contact information into reports generated by task actions this causes tasks that are about unknown contacts to generate reports that are unattached are missing their subject this workflow is used where supervisors with replication depth receive needs signoff reports about patients they can t see to reproduce create a supervisor with replication depth low enough that they don t see patients create a chw under this supervisor configure a form with needs signoff configure a task for the supervisor triggered by the needs signoff form configure this to be resolved if the task action is executed submit the needs signoff form as the chw sync as the supervisor and check if you have generated the task complete the task your task does not resolve it s still in the list and inspecting the resulting report it s missing all patient information expected behavior the resulting report should have the same patient information as the generating report environment instance local app webapp version additional context when written to disk tasks are minified and they only retain a forid property representing action contact id while action contact is deleted when rehydrating the task we rely on the forid contact to exist in order to repopulate action contact we should either a keep action contact when the contact is unknown this allows more information besides a uuid to be injected into the resulting report b use a stub when rehydrating and contact is unknown this means that the resulting report will only have the uuid available and no other information
0
75,656
25,978,379,317
IssuesEvent
2022-12-19 16:36:51
department-of-veterans-affairs/va.gov-cms
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
closed
FE: VAMC System Banner Alert with Situation Updates do not work for Lovell systems
Defect VA.gov frontend ⭐️ Facilities Needs refining
## Describe the defect This issue will need to be refined as work in #11726 is completed. For now this is a placeholder. We simply want to be sure that VAMC System Banner Alert with Situation Updates nodes are displayed properly for Lovell systems... ## To Reproduce Steps to reproduce the behavior: 1. Go to '...' 2. Click on '....' 3. Scroll down to '....' 4. See error ## AC / Expected behavior A clear and concise description of what you expected to happen. ## Screenshots If applicable, add screenshots to help explain your problem. ## Additional context Add any other context about the problem here. Reach out to the Product Managers to determine if it should be escalated as critical (prevents users from accomplishing their work with no known workaround and needs to be addressed within 2 business days). ## Desktop (please complete the following information if relevant, or delete) - OS: [e.g. iOS] - Browser [e.g. chrome, safari] - Version [e.g. 22] ## Labels (You can delete this section once it's complete) - [x] Issue type (red) (defaults to "Defect") - [ ] CMS subsystem (green) - [ ] CMS practice area (blue) - [x] CMS workstream (orange) (not needed for bug tickets) - [ ] CMS-supported product (black) ### CMS Team Please check the team(s) that will do this work. - [ ] `Program` - [ ] `Platform CMS Team` - [ ] `Sitewide Crew` - [ ] `⭐️ Sitewide CMS` - [ ] `⭐️ Public Websites` - [ ] `⭐️ Facilities` - [ ] `⭐️ User support`
1.0
FE: VAMC System Banner Alert with Situation Updates do not work for Lovell systems - ## Describe the defect This issue will need to be refined as work in #11726 is completed. For now this is a placeholder. We simply want to be sure that VAMC System Banner Alert with Situation Updates nodes are displayed properly for Lovell systems... ## To Reproduce Steps to reproduce the behavior: 1. Go to '...' 2. Click on '....' 3. Scroll down to '....' 4. See error ## AC / Expected behavior A clear and concise description of what you expected to happen. ## Screenshots If applicable, add screenshots to help explain your problem. ## Additional context Add any other context about the problem here. Reach out to the Product Managers to determine if it should be escalated as critical (prevents users from accomplishing their work with no known workaround and needs to be addressed within 2 business days). ## Desktop (please complete the following information if relevant, or delete) - OS: [e.g. iOS] - Browser [e.g. chrome, safari] - Version [e.g. 22] ## Labels (You can delete this section once it's complete) - [x] Issue type (red) (defaults to "Defect") - [ ] CMS subsystem (green) - [ ] CMS practice area (blue) - [x] CMS workstream (orange) (not needed for bug tickets) - [ ] CMS-supported product (black) ### CMS Team Please check the team(s) that will do this work. - [ ] `Program` - [ ] `Platform CMS Team` - [ ] `Sitewide Crew` - [ ] `⭐️ Sitewide CMS` - [ ] `⭐️ Public Websites` - [ ] `⭐️ Facilities` - [ ] `⭐️ User support`
defect
fe vamc system banner alert with situation updates do not work for lovell systems describe the defect this issue will need to be refined as work in is completed for now this is a placeholder we simply want to be sure that vamc system banner alert with situation updates nodes are displayed properly for lovell systems to reproduce steps to reproduce the behavior go to click on scroll down to see error ac expected behavior a clear and concise description of what you expected to happen screenshots if applicable add screenshots to help explain your problem additional context add any other context about the problem here reach out to the product managers to determine if it should be escalated as critical prevents users from accomplishing their work with no known workaround and needs to be addressed within business days desktop please complete the following information if relevant or delete os browser version labels you can delete this section once it s complete issue type red defaults to defect cms subsystem green cms practice area blue cms workstream orange not needed for bug tickets cms supported product black cms team please check the team s that will do this work program platform cms team sitewide crew ⭐️ sitewide cms ⭐️ public websites ⭐️ facilities ⭐️ user support
1
177,457
28,494,894,530
IssuesEvent
2023-04-18 13:37:14
rancher/dashboard
https://api.github.com/repos/rancher/dashboard
opened
Kubernetes RBAC competitive analysis
kind/design
We want to understand how other actors deal with the particularities and complexities of Kubernetes RBAC. The initial idea is to compare Rancher's approach with: - K8s Dashboard - Lens - Octant
1.0
Kubernetes RBAC competitive analysis - We want to understand how other actors deal with the particularities and complexities of Kubernetes RBAC. The initial idea is to compare Rancher's approach with: - K8s Dashboard - Lens - Octant
non_defect
kubernetes rbac competitive analysis we want to understand how other actors deal with the particularities and complexities of kubernetes rbac the initial idea is to compare rancher s approach with dashboard lens octant
0
493,417
14,231,702,816
IssuesEvent
2020-11-18 09:55:51
kubermatic/kubermatic
https://api.github.com/repos/kubermatic/kubermatic
closed
Support MachineDeployment specific kubelet config
customer-request lifecycle/rotten priority/low team/lifecycle
**User Story** The kubelet configuration is currently applied as a ConfigMap from the addons image. Customer wants to change the API request limits per cluster or even MachineDeployment. E.g., different limits for small and large nodes. **Acceptance criteria** Kubelet configuration should be adaptable per MachineDeployment
1.0
Support MachineDeployment specific kubelet config - **User Story** The kubelet configuration is currently applied as a ConfigMap from the addons image. Customer wants to change the API request limits per cluster or even MachineDeployment. E.g., different limits for small and large nodes. **Acceptance criteria** Kubelet configuration should be adaptable per MachineDeployment
non_defect
support machinedeployment specific kubelet config user story the kubelet configuration is currently applied as a configmap from the addons image customer wants to change the api request limits per cluster or even machinedeployment e g different limits for small and large nodes acceptance criteria kubelet configuration should be adaptable per machinedeployment
0
542,929
15,874,406,843
IssuesEvent
2021-04-09 04:57:34
remnoteio/remnote-issues
https://api.github.com/repos/remnoteio/remnote-issues
closed
No. of Tagged Rems does not appear or appear incorrectly at bottom portal results
priority=2
Although there is indication of tagged rems existing in the bottom portal, upon expanding the portal there would be no results shown. In other instances, an incorrect count of Tagged Rems is reflected. --- ![Screenshot 2021-02-22 at 2 11 20 PM](https://user-images.githubusercontent.com/10230857/108669713-866eb800-7518-11eb-96da-635bb738bb1d.png) --- ![Screenshot 2021-02-22 at 2 11 32 PM](https://user-images.githubusercontent.com/10230857/108669721-8a023f00-7518-11eb-8eb8-023254accc9e.png) --- ![Screenshot 2021-02-22 at 2 26 00 PM](https://user-images.githubusercontent.com/10230857/108670571-0ba69c80-751a-11eb-9445-89f99bfb8c80.png) --- ![Screenshot 2021-02-22 at 2 26 09 PM](https://user-images.githubusercontent.com/10230857/108670587-12351400-751a-11eb-916e-e02ed640718e.png)
1.0
No. of Tagged Rems does not appear or appear incorrectly at bottom portal results - Although there is indication of tagged rems existing in the bottom portal, upon expanding the portal there would be no results shown. In other instances, an incorrect count of Tagged Rems is reflected. --- ![Screenshot 2021-02-22 at 2 11 20 PM](https://user-images.githubusercontent.com/10230857/108669713-866eb800-7518-11eb-96da-635bb738bb1d.png) --- ![Screenshot 2021-02-22 at 2 11 32 PM](https://user-images.githubusercontent.com/10230857/108669721-8a023f00-7518-11eb-8eb8-023254accc9e.png) --- ![Screenshot 2021-02-22 at 2 26 00 PM](https://user-images.githubusercontent.com/10230857/108670571-0ba69c80-751a-11eb-9445-89f99bfb8c80.png) --- ![Screenshot 2021-02-22 at 2 26 09 PM](https://user-images.githubusercontent.com/10230857/108670587-12351400-751a-11eb-916e-e02ed640718e.png)
non_defect
no of tagged rems does not appear or appear incorrectly at bottom portal results although there is indication of tagged rems existing in the bottom portal upon expanding the portal there would be no results shown in other instances an incorrect count of tagged rems is reflected
0
78,768
27,751,928,124
IssuesEvent
2023-03-15 21:31:41
jurgendl/hql-builder
https://api.github.com/repos/jurgendl/hql-builder
closed
remove HikariCP, fallback to previous connection-pool
bug Type-Defect Priority-Critical no-issue-activity
remove HikariCP, fallback to previous connection-pool HikariCP seems to hang
1.0
remove HikariCP, fallback to previous connection-pool - remove HikariCP, fallback to previous connection-pool HikariCP seems to hang
defect
remove hikaricp fallback to previous connection pool remove hikaricp fallback to previous connection pool hikaricp seems to hang
1
67,673
8,170,283,302
IssuesEvent
2018-08-27 07:04:35
otavanopisto/muikku
https://api.github.com/repos/otavanopisto/muikku
closed
Audio Component
MATERIALS REDESIGN2017 enhancement
_From @onzag on August 27, 2018 6:3_ null _Copied from original issue: otavanopisto/board#1031_
1.0
Audio Component - _From @onzag on August 27, 2018 6:3_ null _Copied from original issue: otavanopisto/board#1031_
non_defect
audio component from onzag on august null copied from original issue otavanopisto board
0
75,397
25,810,854,823
IssuesEvent
2022-12-11 20:56:38
DependencyTrack/dependency-track
https://api.github.com/repos/DependencyTrack/dependency-track
closed
Dependency track takes more than 30 minutes to analyze and Dependency tracker jenkins aborts due to polling time limti.
defect
The defect may already be reported! Please search for the defect before creating one. ### Current Behavior: Jenkins Dependecy tarck jobs fails since server takes more 30 minutes analyze the project. ### Steps to Reproduce: Run jenkins Dependecy track job in Jenkins pipeline for java project which has more than 1000 components(These are java modules) ### Expected Behavior: Jenkins PipleLine should not fail. ### Environment: - Dependency-Track Version: 'latest' - Distribution: [ Docker ] - BOM Format & Version: "bomFormat": "CycloneDX", "specVersion": "1.4", "version": 1, - Database Server: PostgreSQL - Browser: Chrome ### Additional Details: 22-06-28 06:37:51,463 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed portfolio metrics update 2022-06-28 06:37:56,340 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed metrics update for project: cd6feaf1-c6df-4c15-a7b4-f118151a6605 2022-06-28 06:52:29,046 [] INFO [org.dependencytrack.tasks.BomUploadProcessingTask] Processing CycloneDX BOM uploaded to project: cd6feaf1-c6df-4c15-a7b4-f118151a6605 2022-06-28 06:56:59,202 [] INFO [org.dependencytrack.tasks.BomUploadProcessingTask] Processing CycloneDX dependency graph for project: cd6feaf1-c6df-4c15-a7b4-f118151a6605 2022-06-28 07:04:14,351 [] INFO [org.dependencytrack.tasks.BomUploadProcessingTask] Processed 18477 components and 0 services uploaded to project cd6feaf1-c6df-4c15-a7b4-f118151a6605 2022-06-28 07:22:08,107 [] INFO [org.dependencytrack.tasks.scanners.InternalAnalysisTask] Starting internal analysis task 2022-06-28 07:22:42,573 [] INFO [org.dependencytrack.tasks.scanners.InternalAnalysisTask] Internal analysis complete 2022-06-28 07:22:42,579 [] WARN [org.dependencytrack.tasks.scanners.OssIndexAnalysisTask] An API username or token has not been specified for use with OSS Index. Using anonymous access 2022-06-28 07:22:42,579 [] INFO [org.dependencytrack.tasks.scanners.OssIndexAnalysisTask] Starting Sonatype OSS Index analysis task 2022-06-28 07:23:21,187 [] INFO [org.dependencytrack.tasks.scanners.OssIndexAnalysisTask] Sonatype OSS Index analysis complete 2022-06-28 07:23:21,189 [] INFO [org.dependencytrack.policy.PolicyEngine] Evaluating 18477 component(s) against applicable policies 2022-06-28 07:24:34,036 [] INFO [org.dependencytrack.policy.PolicyEngine] Policy analysis complete 2022-06-28 07:24:34,038 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing metrics update for project: cd6feaf1-c6df-4c15-a7b4-f118151a6605 2022-06-28 07:24:41,373 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed metrics update for project: cd6feaf1-c6df-4c15-a7b4-f118151a6605 2022-06-28 07:25:37,850 [] INFO [org.dependencytrack.tasks.ClearComponentAnalysisCacheTask] Clearing ComponentAnalysisCache 2022-06-28 07:25:37,909 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing metrics update on vulnerability database 2022-06-28 07:25:37,933 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing portfolio metrics update 2022-06-28 07:25:37,945 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing metrics update for project: 7be7ded7-ac02-40c1-a3f8-ca11d1bc005f 2022-06-28 07:25:40,410 [] INFO [org.dependencytrack.tasks.ClearComponentAnalysisCacheTask] Complete 2022-06-28 07:25:46,910 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed metrics update for project: 7be7ded7-ac02-40c1-a3f8-ca11d1bc005f 2022-06-28 07:25:46,914 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing metrics update for project: cd6feaf1-c6df-4c15-a7b4-f118151a6605 2022-06-28 07:25:55,015 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed metrics update for project: cd6feaf1-c6df-4c15-a7b4-f118151a6605 2022-06-28 07:25:55,030 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed portfolio metrics update 2022-06-28 07:26:27,843 [] INFO [org.dependencytrack.tasks.VulnDbSyncTask] Starting VulnDB mirror synchronization task 2022-06-28 07:26:27,844 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Starting NIST mirroring task 2022-06-28 07:26:27,844 [] INFO [org.dependencytrack.tasks.VulnDbSyncTask] VulnDB mirror directory does not exist. Skipping. 2022-06-28 07:26:27,848 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Downloading files at Tue Jun 28 07:26:27 UTC 2022 2022-06-28 07:26:28,618 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Initiating download of https://nvd.nist.gov/feeds/json/cve/1.1/nvdcve-1.1-modified.json.gz 2022-06-28 07:26:28,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Downloading... 2022-06-28 07:26:29,602 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Uncompressing nvdcve-1.1-modified.json.gz 2022-06-28 07:26:29,711 [] INFO [org.dependencytrack.parser.nvd.NvdParser] Parsing nvdcve-1.1-modified.json 2022-06-28 07:30:44,409 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed metrics update on vulnerability database 2022-06-28 07:33:25,101 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Initiating download of https://nvd.nist.gov/feeds/json/cve/1.1/nvdcve-1.1-modified.meta 2022-06-28 07:33:25,964 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Downloading... 2022-06-28 07:33:25,972 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2022.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,972 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2021.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,972 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2020.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,972 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2019.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,972 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2018.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,973 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2017.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,973 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2016.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,973 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2015.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,973 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2014.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,973 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2013.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,973 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2012.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,973 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2011.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2010.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2009.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2008.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2007.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2006.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2005.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2004.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2003.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2002.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,975 [] INFO [org.dependencytrack.tasks.NistMirrorTask] NIST mirroring complete 2022-06-28 07:33:25,975 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Time spent (d/l): 1219ms 2022-06-28 07:33:25,975 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Time spent (parse): 415389ms 2022-06-28 07:33:25,975 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Time spent (total): 418131ms 2022-06-28 07:33:26,466 [] INFO [org.dependencytrack.tasks.EpssMirrorTask] Starting EPSS mirroring task 2022-06-28 07:33:26,466 [] INFO [org.dependencytrack.tasks.EpssMirrorTask] Initiating download of https://epss.cyentia.com/epss_scores-current.csv.gz 2022-06-28 07:33:27,918 [] INFO [org.dependencytrack.tasks.EpssMirrorTask] Downloading... 2022-06-28 07:33:28,448 [] INFO [org.dependencytrack.tasks.EpssMirrorTask] Uncompressing epss_scores-current.csv.gz 2022-06-28 07:33:28,533 [] INFO [org.dependencytrack.parser.epss.EpssParser] Parsing epss_scores-current.csv 2022-06-28 07:52:55,034 [] INFO [org.dependencytrack.tasks.EpssMirrorTask] EPSS mirroring complete 2022-06-28 07:52:55,034 [] INFO [org.dependencytrack.tasks.EpssMirrorTask] Time spent (d/l): 1452ms 2022-06-28 07:52:55,034 [] INFO [org.dependencytrack.tasks.EpssMirrorTask] Time spent (parse): 1166500ms 2022-06-28 07:52:55,034 [] INFO [org.dependencytrack.tasks.EpssMirrorTask] Time spent (total): 1168568ms 2022-06-28 08:25:27,852 [] INFO [org.dependencytrack.tasks.InternalComponentIdentificationTask] Starting internal component identification task 2022-06-28 08:25:27,914 [] INFO [org.dependencytrack.tasks.repositories.RepositoryMetaAnalyzerTask] Performing component repository metadata analysis against 564 components in project: 7be7ded7-ac02-40c1-a3f8-ca11d1bc005f 2022-06-28 08:25:29,630 [] INFO [org.dependencytrack.tasks.InternalComponentIdentificationTask] Internal component identification task completed 2022-06-28 08:25:37,909 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing metrics update on vulnerability database 2022-06-28 08:25:37,934 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing portfolio metrics update 2022-06-28 08:25:37,942 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing metrics update for project: 7be7ded7-ac02-40c1-a3f8-ca11d1bc005f 2022-06-28 08:25:48,722 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed metrics update for project: 7be7ded7-ac02-40c1-a3f8-ca11d1bc005f 2022-06-28 08:25:48,725 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing metrics update for project: cd6feaf1-c6df-4c15-a7b4-f118151a6605 2022-06-28 08:25:57,114 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed metrics update for project: cd6feaf1-c6df-4c15-a7b4-f118151a6605 2022-06-28 08:25:57,132 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed portfolio metrics update 2022-06-28 08:30:36,552 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed metrics update on vulnerability database 2022-06-28 08:34:26,719 [] INFO [org.dependencytrack.tasks.repositories.RepositoryMetaAnalyzerTask] Completed component repository metadata analysis against 564 components in project: 7be7ded7-ac02-40c1-a3f8-ca11d1bc005f 2022-06-28 08:34:26,757 [] INFO [org.dependencytrack.tasks.repositories.RepositoryMetaAnalyzerTask] Performing component repository metadata analysis against 564 components in project: cd6feaf1-c6df-4c15-a7b4-f118151a6605 2022-06-28 08:42:17,083 [] INFO [org.dependencytrack.tasks.repositories.RepositoryMetaAnalyzerTask] Completed component repository metadata analysis against 564 components in project: cd6feaf1-c6df-4c15-a7b4-f118151a6605 2022-06-28 08:42:17,086 [] INFO [org.dependencytrack.tasks.repositories.RepositoryMetaAnalyzerTask] Portfolio component repository metadata analysis complete dtrack@159c5a33cb10:/data$
1.0
Dependency track takes more than 30 minutes to analyze and Dependency tracker jenkins aborts due to polling time limti. - The defect may already be reported! Please search for the defect before creating one. ### Current Behavior: Jenkins Dependecy tarck jobs fails since server takes more 30 minutes analyze the project. ### Steps to Reproduce: Run jenkins Dependecy track job in Jenkins pipeline for java project which has more than 1000 components(These are java modules) ### Expected Behavior: Jenkins PipleLine should not fail. ### Environment: - Dependency-Track Version: 'latest' - Distribution: [ Docker ] - BOM Format & Version: "bomFormat": "CycloneDX", "specVersion": "1.4", "version": 1, - Database Server: PostgreSQL - Browser: Chrome ### Additional Details: 22-06-28 06:37:51,463 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed portfolio metrics update 2022-06-28 06:37:56,340 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed metrics update for project: cd6feaf1-c6df-4c15-a7b4-f118151a6605 2022-06-28 06:52:29,046 [] INFO [org.dependencytrack.tasks.BomUploadProcessingTask] Processing CycloneDX BOM uploaded to project: cd6feaf1-c6df-4c15-a7b4-f118151a6605 2022-06-28 06:56:59,202 [] INFO [org.dependencytrack.tasks.BomUploadProcessingTask] Processing CycloneDX dependency graph for project: cd6feaf1-c6df-4c15-a7b4-f118151a6605 2022-06-28 07:04:14,351 [] INFO [org.dependencytrack.tasks.BomUploadProcessingTask] Processed 18477 components and 0 services uploaded to project cd6feaf1-c6df-4c15-a7b4-f118151a6605 2022-06-28 07:22:08,107 [] INFO [org.dependencytrack.tasks.scanners.InternalAnalysisTask] Starting internal analysis task 2022-06-28 07:22:42,573 [] INFO [org.dependencytrack.tasks.scanners.InternalAnalysisTask] Internal analysis complete 2022-06-28 07:22:42,579 [] WARN [org.dependencytrack.tasks.scanners.OssIndexAnalysisTask] An API username or token has not been specified for use with OSS Index. Using anonymous access 2022-06-28 07:22:42,579 [] INFO [org.dependencytrack.tasks.scanners.OssIndexAnalysisTask] Starting Sonatype OSS Index analysis task 2022-06-28 07:23:21,187 [] INFO [org.dependencytrack.tasks.scanners.OssIndexAnalysisTask] Sonatype OSS Index analysis complete 2022-06-28 07:23:21,189 [] INFO [org.dependencytrack.policy.PolicyEngine] Evaluating 18477 component(s) against applicable policies 2022-06-28 07:24:34,036 [] INFO [org.dependencytrack.policy.PolicyEngine] Policy analysis complete 2022-06-28 07:24:34,038 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing metrics update for project: cd6feaf1-c6df-4c15-a7b4-f118151a6605 2022-06-28 07:24:41,373 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed metrics update for project: cd6feaf1-c6df-4c15-a7b4-f118151a6605 2022-06-28 07:25:37,850 [] INFO [org.dependencytrack.tasks.ClearComponentAnalysisCacheTask] Clearing ComponentAnalysisCache 2022-06-28 07:25:37,909 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing metrics update on vulnerability database 2022-06-28 07:25:37,933 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing portfolio metrics update 2022-06-28 07:25:37,945 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing metrics update for project: 7be7ded7-ac02-40c1-a3f8-ca11d1bc005f 2022-06-28 07:25:40,410 [] INFO [org.dependencytrack.tasks.ClearComponentAnalysisCacheTask] Complete 2022-06-28 07:25:46,910 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed metrics update for project: 7be7ded7-ac02-40c1-a3f8-ca11d1bc005f 2022-06-28 07:25:46,914 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing metrics update for project: cd6feaf1-c6df-4c15-a7b4-f118151a6605 2022-06-28 07:25:55,015 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed metrics update for project: cd6feaf1-c6df-4c15-a7b4-f118151a6605 2022-06-28 07:25:55,030 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed portfolio metrics update 2022-06-28 07:26:27,843 [] INFO [org.dependencytrack.tasks.VulnDbSyncTask] Starting VulnDB mirror synchronization task 2022-06-28 07:26:27,844 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Starting NIST mirroring task 2022-06-28 07:26:27,844 [] INFO [org.dependencytrack.tasks.VulnDbSyncTask] VulnDB mirror directory does not exist. Skipping. 2022-06-28 07:26:27,848 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Downloading files at Tue Jun 28 07:26:27 UTC 2022 2022-06-28 07:26:28,618 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Initiating download of https://nvd.nist.gov/feeds/json/cve/1.1/nvdcve-1.1-modified.json.gz 2022-06-28 07:26:28,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Downloading... 2022-06-28 07:26:29,602 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Uncompressing nvdcve-1.1-modified.json.gz 2022-06-28 07:26:29,711 [] INFO [org.dependencytrack.parser.nvd.NvdParser] Parsing nvdcve-1.1-modified.json 2022-06-28 07:30:44,409 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed metrics update on vulnerability database 2022-06-28 07:33:25,101 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Initiating download of https://nvd.nist.gov/feeds/json/cve/1.1/nvdcve-1.1-modified.meta 2022-06-28 07:33:25,964 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Downloading... 2022-06-28 07:33:25,972 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2022.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,972 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2021.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,972 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2020.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,972 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2019.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,972 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2018.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,973 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2017.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,973 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2016.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,973 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2015.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,973 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2014.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,973 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2013.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,973 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2012.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,973 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2011.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2010.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2009.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2008.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2007.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2006.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2005.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2004.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2003.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2002.json.gz not necessary. Will use modified feed for updates. 2022-06-28 07:33:25,975 [] INFO [org.dependencytrack.tasks.NistMirrorTask] NIST mirroring complete 2022-06-28 07:33:25,975 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Time spent (d/l): 1219ms 2022-06-28 07:33:25,975 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Time spent (parse): 415389ms 2022-06-28 07:33:25,975 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Time spent (total): 418131ms 2022-06-28 07:33:26,466 [] INFO [org.dependencytrack.tasks.EpssMirrorTask] Starting EPSS mirroring task 2022-06-28 07:33:26,466 [] INFO [org.dependencytrack.tasks.EpssMirrorTask] Initiating download of https://epss.cyentia.com/epss_scores-current.csv.gz 2022-06-28 07:33:27,918 [] INFO [org.dependencytrack.tasks.EpssMirrorTask] Downloading... 2022-06-28 07:33:28,448 [] INFO [org.dependencytrack.tasks.EpssMirrorTask] Uncompressing epss_scores-current.csv.gz 2022-06-28 07:33:28,533 [] INFO [org.dependencytrack.parser.epss.EpssParser] Parsing epss_scores-current.csv 2022-06-28 07:52:55,034 [] INFO [org.dependencytrack.tasks.EpssMirrorTask] EPSS mirroring complete 2022-06-28 07:52:55,034 [] INFO [org.dependencytrack.tasks.EpssMirrorTask] Time spent (d/l): 1452ms 2022-06-28 07:52:55,034 [] INFO [org.dependencytrack.tasks.EpssMirrorTask] Time spent (parse): 1166500ms 2022-06-28 07:52:55,034 [] INFO [org.dependencytrack.tasks.EpssMirrorTask] Time spent (total): 1168568ms 2022-06-28 08:25:27,852 [] INFO [org.dependencytrack.tasks.InternalComponentIdentificationTask] Starting internal component identification task 2022-06-28 08:25:27,914 [] INFO [org.dependencytrack.tasks.repositories.RepositoryMetaAnalyzerTask] Performing component repository metadata analysis against 564 components in project: 7be7ded7-ac02-40c1-a3f8-ca11d1bc005f 2022-06-28 08:25:29,630 [] INFO [org.dependencytrack.tasks.InternalComponentIdentificationTask] Internal component identification task completed 2022-06-28 08:25:37,909 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing metrics update on vulnerability database 2022-06-28 08:25:37,934 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing portfolio metrics update 2022-06-28 08:25:37,942 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing metrics update for project: 7be7ded7-ac02-40c1-a3f8-ca11d1bc005f 2022-06-28 08:25:48,722 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed metrics update for project: 7be7ded7-ac02-40c1-a3f8-ca11d1bc005f 2022-06-28 08:25:48,725 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing metrics update for project: cd6feaf1-c6df-4c15-a7b4-f118151a6605 2022-06-28 08:25:57,114 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed metrics update for project: cd6feaf1-c6df-4c15-a7b4-f118151a6605 2022-06-28 08:25:57,132 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed portfolio metrics update 2022-06-28 08:30:36,552 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed metrics update on vulnerability database 2022-06-28 08:34:26,719 [] INFO [org.dependencytrack.tasks.repositories.RepositoryMetaAnalyzerTask] Completed component repository metadata analysis against 564 components in project: 7be7ded7-ac02-40c1-a3f8-ca11d1bc005f 2022-06-28 08:34:26,757 [] INFO [org.dependencytrack.tasks.repositories.RepositoryMetaAnalyzerTask] Performing component repository metadata analysis against 564 components in project: cd6feaf1-c6df-4c15-a7b4-f118151a6605 2022-06-28 08:42:17,083 [] INFO [org.dependencytrack.tasks.repositories.RepositoryMetaAnalyzerTask] Completed component repository metadata analysis against 564 components in project: cd6feaf1-c6df-4c15-a7b4-f118151a6605 2022-06-28 08:42:17,086 [] INFO [org.dependencytrack.tasks.repositories.RepositoryMetaAnalyzerTask] Portfolio component repository metadata analysis complete dtrack@159c5a33cb10:/data$
defect
dependency track takes more than minutes to analyze and dependency tracker jenkins aborts due to polling time limti the defect may already be reported please search for the defect before creating one current behavior jenkins dependecy tarck jobs fails since server takes more minutes analyze the project steps to reproduce run jenkins dependecy track job in jenkins pipeline for java project which has more than components these are java modules expected behavior jenkins pipleline should not fail environment dependency track version latest distribution bom format version bomformat cyclonedx specversion version database server postgresql browser chrome additional details info completed portfolio metrics update info completed metrics update for project info processing cyclonedx bom uploaded to project info processing cyclonedx dependency graph for project info processed components and services uploaded to project info starting internal analysis task info internal analysis complete warn an api username or token has not been specified for use with oss index using anonymous access info starting sonatype oss index analysis task info sonatype oss index analysis complete info evaluating component s against applicable policies info policy analysis complete info executing metrics update for project info completed metrics update for project info clearing componentanalysiscache info executing metrics update on vulnerability database info executing portfolio metrics update info executing metrics update for project info complete info completed metrics update for project info executing metrics update for project info completed metrics update for project info completed portfolio metrics update info starting vulndb mirror synchronization task info starting nist mirroring task info vulndb mirror directory does not exist skipping info downloading files at tue jun utc info initiating download of info downloading info uncompressing nvdcve modified json gz info parsing nvdcve modified json info completed metrics update on vulnerability database info initiating download of info downloading info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info nist mirroring complete info time spent d l info time spent parse info time spent total info starting epss mirroring task info initiating download of info downloading info uncompressing epss scores current csv gz info parsing epss scores current csv info epss mirroring complete info time spent d l info time spent parse info time spent total info starting internal component identification task info performing component repository metadata analysis against components in project info internal component identification task completed info executing metrics update on vulnerability database info executing portfolio metrics update info executing metrics update for project info completed metrics update for project info executing metrics update for project info completed metrics update for project info completed portfolio metrics update info completed metrics update on vulnerability database info completed component repository metadata analysis against components in project info performing component repository metadata analysis against components in project info completed component repository metadata analysis against components in project info portfolio component repository metadata analysis complete dtrack data
1
566,673
16,826,842,335
IssuesEvent
2021-06-17 19:50:22
rokwire/illinois-app
https://api.github.com/repos/rokwire/illinois-app
opened
[BUG] Career center link is not working
Priority: High Type: Bug
**Describe the bug** Student Guide - The Career center link is not doing anything when the user taps on it. **To Reproduce** Steps to reproduce the behavior: 1. Install the Illinois app 2. Complete the onboarding and verify the role as a student role 3. Via browse screen, navigate to Student guide and navigate to Career -> Get Ready for the job Market 4. Tap on Career Center. **Actual Screen** The Career center link is not doing anything when the user taps on it. **Expected behavior** The Career center webpage should load when the user tap on the link* **Screenshots** ![career c](https://user-images.githubusercontent.com/26231179/122463375-27c3c300-cf7b-11eb-84ea-72f0df48eb5c.png) ![CC1](https://user-images.githubusercontent.com/26231179/122463437-38743900-cf7b-11eb-9a40-a25b864e1d48.png) **Smartphone (please complete the following information):** - Device: [e.g. Android] - Version [e.g. 2.4.25]
1.0
[BUG] Career center link is not working - **Describe the bug** Student Guide - The Career center link is not doing anything when the user taps on it. **To Reproduce** Steps to reproduce the behavior: 1. Install the Illinois app 2. Complete the onboarding and verify the role as a student role 3. Via browse screen, navigate to Student guide and navigate to Career -> Get Ready for the job Market 4. Tap on Career Center. **Actual Screen** The Career center link is not doing anything when the user taps on it. **Expected behavior** The Career center webpage should load when the user tap on the link* **Screenshots** ![career c](https://user-images.githubusercontent.com/26231179/122463375-27c3c300-cf7b-11eb-84ea-72f0df48eb5c.png) ![CC1](https://user-images.githubusercontent.com/26231179/122463437-38743900-cf7b-11eb-9a40-a25b864e1d48.png) **Smartphone (please complete the following information):** - Device: [e.g. Android] - Version [e.g. 2.4.25]
non_defect
career center link is not working describe the bug student guide the career center link is not doing anything when the user taps on it to reproduce steps to reproduce the behavior install the illinois app complete the onboarding and verify the role as a student role via browse screen navigate to student guide and navigate to career get ready for the job market tap on career center actual screen the career center link is not doing anything when the user taps on it expected behavior the career center webpage should load when the user tap on the link screenshots smartphone please complete the following information device version
0
5,674
2,610,193,081
IssuesEvent
2015-02-26 19:00:58
chrsmith/quchuseban
https://api.github.com/repos/chrsmith/quchuseban
opened
推荐脸上长色斑去除偏方
auto-migrated Priority-Medium Type-Defect
``` 《摘要》 四季的轮回,变换着多样缤纷的色彩。当黄,静静的铺满山坡。一丝怅然恐慌成无法言说的失落。凝不住经年的妩媚,留不下往昔的绚烂。在季节的深处,九月的风微凉着心事。一场寒雨在风中呜咽叹息,轻敲轩窗,碎断人肠。没有了昨日纤雨湿花的雅致,忘却了曾经柳絮扑帘的悠然。绮梦悠悠伤情绵绵,一任愁绪在风中雨里飘摇成伤,恍若,枝头颤动的叶子,没了昨日张扬的碧绿,沉淀了一个春夏的水嫩在风中瞬间走失,只留下一袭苍白,挣扎在季节的末梢。嘴角溢出的笑,潋滟了一湖的秋水。如果出去嘴角的色斑,我想这个微笑会更美.脸上长色斑去除偏方, 《客户案例》   莫小姐 24岁<br>   长时间面对电脑的坏处我一直都知道,像眼睛疲劳、干�� �,身体处于亚健康之类的,工作一年来我基本上10个小时对�� �电脑,好在我什么不舒服的症状都没出现,正当我庆幸不已� ��同时,我发现我额头、鼻梁两边和上眼睑处长斑了,数量很 多,就像雨后春笋一样,一点一点在短时间内全部冒出来了�� �看着脸上那如同苍蝇屎大小的斑点,颜色为黄色,略比肤色� ��点,远看近看都明显。哪个女孩子不爱美?为此特别烦恼。<b r>   在网上搜索祛斑产品时发现了「黛芙薇尔精华液」,了�� �了一番之后觉得还不错,而且网上论坛里面大家发帖子都说� ��黛芙薇尔精华液」好,没发现有说副作用的,然后我还上「 黛芙薇尔精华液」的网站很认真的看了一下产品说明,说明�� �字很专业详细,对治疗机理、产品优势描述得挺不错的,我� ��在网上订购了一个周期的。<br>   大概了十几天的样子,我鼻梁两边的斑块颜色就变淡了�� �面积也慢慢在缩小,本来粗糙、多油的皮肤状况也得到了改� ��。一个月过去以后,对着镜子看自己脸上的斑,鼻梁上的大 斑块已经缩小了一大半,颜色由以前的深褐色变成了浅褐色�� �额头和眼角处的斑也变淡变浅了。「黛芙薇尔精华液」祛斑� ��家在我订购产品的时候就说过,产品在祛斑的同时还能调经 ,让经期变得正常,还能减轻痛经的痛苦。我使用第二个周�� �产品的时候,这方面的效果显示出来了,我以前痛经非常厉� ��,自从使用这个套装以来,不但月经量变得正常了,而且痛 经也没那么严重了,在祛斑的同时还能调经,「黛芙薇尔精�� �液」真的很不错。 很快我就使用完第二个周期的产品,鼻梁两边的斑基本上看�� �到了,颜色跟肤色很接近,不仔细看根本看不出来了,我可� ��不用粉底盖住那一块的斑。额头上的斑也在慢慢消退,特别 是眼睑处的斑,淡化的比较明显。我接着使用第三个周期的�� �品,这个周期见效很快,刚半个月多一点,眼睑处的斑就消� ��了,褐色的斑融进皮肤里面,一点都看不出来了,额头上的 也差不多没有了。使用完三个周期的「黛芙薇尔精华液」之�� �,我从斑女人变成了一个水嫩白净的女人,而且我的月经也� ��得很有规律,内分泌正常,心情也变得非常开朗,哪位想要 祛斑的朋友就学我的祛斑方法,相信通过使用「黛芙薇尔精�� �液」,大家都能祛斑成功。 阅读了脸上长色斑去除偏方,再看脸上容易长斑的原因: 《色斑形成原因》   内部因素   一、压力   当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。   二、荷尔蒙分泌失调   避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。   三、新陈代谢缓慢   肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。   四、错误的使用化妆品   使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。   外部因素   一、紫外线   照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。   二、不良的清洁习惯   因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。   三、遗传基因   父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》   1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐�� �去掉吗?   答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新�� �客都是通过老顾客介绍而来,口碑由此而来!   2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?   答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技�� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!   3,去除黄褐斑之后,会反弹吗?   答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌!我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗?   4,你们的价格有点贵,能不能便宜一点?   答:如果您使用西药最少需要2000元,煎服的药最少需要3 000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗   5,我适合用黛芙薇尔精华液吗?   答:黛芙薇尔适用人群:   1、生理紊乱引起的黄褐斑人群   2、生育引起的妊娠斑人群   3、年纪增长引起的老年斑人群   4、化妆品色素沉积、辐射斑人群   5、长期日照引起的日晒斑人群   6、肌肤暗淡急需美白的人群 《祛斑小方法》 脸上长色斑去除偏方,同时为您分享祛斑小方法 一本中医古书上看到了一则去除雀斑的方子:将黄豆(黄豆�� �生的)浸泡在醋中一个月,每天服用几粒,坚持一段时间即� ��完全消除。 (去斑讲究内外兼修,这点值得一试) 。 ``` ----- Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 5:38
1.0
推荐脸上长色斑去除偏方 - ``` 《摘要》 四季的轮回,变换着多样缤纷的色彩。当黄,静静的铺满山坡。一丝怅然恐慌成无法言说的失落。凝不住经年的妩媚,留不下往昔的绚烂。在季节的深处,九月的风微凉着心事。一场寒雨在风中呜咽叹息,轻敲轩窗,碎断人肠。没有了昨日纤雨湿花的雅致,忘却了曾经柳絮扑帘的悠然。绮梦悠悠伤情绵绵,一任愁绪在风中雨里飘摇成伤,恍若,枝头颤动的叶子,没了昨日张扬的碧绿,沉淀了一个春夏的水嫩在风中瞬间走失,只留下一袭苍白,挣扎在季节的末梢。嘴角溢出的笑,潋滟了一湖的秋水。如果出去嘴角的色斑,我想这个微笑会更美.脸上长色斑去除偏方, 《客户案例》   莫小姐 24岁<br>   长时间面对电脑的坏处我一直都知道,像眼睛疲劳、干�� �,身体处于亚健康之类的,工作一年来我基本上10个小时对�� �电脑,好在我什么不舒服的症状都没出现,正当我庆幸不已� ��同时,我发现我额头、鼻梁两边和上眼睑处长斑了,数量很 多,就像雨后春笋一样,一点一点在短时间内全部冒出来了�� �看着脸上那如同苍蝇屎大小的斑点,颜色为黄色,略比肤色� ��点,远看近看都明显。哪个女孩子不爱美?为此特别烦恼。<b r>   在网上搜索祛斑产品时发现了「黛芙薇尔精华液」,了�� �了一番之后觉得还不错,而且网上论坛里面大家发帖子都说� ��黛芙薇尔精华液」好,没发现有说副作用的,然后我还上「 黛芙薇尔精华液」的网站很认真的看了一下产品说明,说明�� �字很专业详细,对治疗机理、产品优势描述得挺不错的,我� ��在网上订购了一个周期的。<br>   大概了十几天的样子,我鼻梁两边的斑块颜色就变淡了�� �面积也慢慢在缩小,本来粗糙、多油的皮肤状况也得到了改� ��。一个月过去以后,对着镜子看自己脸上的斑,鼻梁上的大 斑块已经缩小了一大半,颜色由以前的深褐色变成了浅褐色�� �额头和眼角处的斑也变淡变浅了。「黛芙薇尔精华液」祛斑� ��家在我订购产品的时候就说过,产品在祛斑的同时还能调经 ,让经期变得正常,还能减轻痛经的痛苦。我使用第二个周�� �产品的时候,这方面的效果显示出来了,我以前痛经非常厉� ��,自从使用这个套装以来,不但月经量变得正常了,而且痛 经也没那么严重了,在祛斑的同时还能调经,「黛芙薇尔精�� �液」真的很不错。 很快我就使用完第二个周期的产品,鼻梁两边的斑基本上看�� �到了,颜色跟肤色很接近,不仔细看根本看不出来了,我可� ��不用粉底盖住那一块的斑。额头上的斑也在慢慢消退,特别 是眼睑处的斑,淡化的比较明显。我接着使用第三个周期的�� �品,这个周期见效很快,刚半个月多一点,眼睑处的斑就消� ��了,褐色的斑融进皮肤里面,一点都看不出来了,额头上的 也差不多没有了。使用完三个周期的「黛芙薇尔精华液」之�� �,我从斑女人变成了一个水嫩白净的女人,而且我的月经也� ��得很有规律,内分泌正常,心情也变得非常开朗,哪位想要 祛斑的朋友就学我的祛斑方法,相信通过使用「黛芙薇尔精�� �液」,大家都能祛斑成功。 阅读了脸上长色斑去除偏方,再看脸上容易长斑的原因: 《色斑形成原因》   内部因素   一、压力   当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。   二、荷尔蒙分泌失调   避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。   三、新陈代谢缓慢   肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。   四、错误的使用化妆品   使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。   外部因素   一、紫外线   照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。   二、不良的清洁习惯   因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。   三、遗传基因   父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》   1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐�� �去掉吗?   答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新�� �客都是通过老顾客介绍而来,口碑由此而来!   2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?   答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技�� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!   3,去除黄褐斑之后,会反弹吗?   答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌!我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗?   4,你们的价格有点贵,能不能便宜一点?   答:如果您使用西药最少需要2000元,煎服的药最少需要3 000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗   5,我适合用黛芙薇尔精华液吗?   答:黛芙薇尔适用人群:   1、生理紊乱引起的黄褐斑人群   2、生育引起的妊娠斑人群   3、年纪增长引起的老年斑人群   4、化妆品色素沉积、辐射斑人群   5、长期日照引起的日晒斑人群   6、肌肤暗淡急需美白的人群 《祛斑小方法》 脸上长色斑去除偏方,同时为您分享祛斑小方法 一本中医古书上看到了一则去除雀斑的方子:将黄豆(黄豆�� �生的)浸泡在醋中一个月,每天服用几粒,坚持一段时间即� ��完全消除。 (去斑讲究内外兼修,这点值得一试) 。 ``` ----- Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 5:38
defect
推荐脸上长色斑去除偏方 《摘要》 四季的轮回,变换着多样缤纷的色彩。当黄,静静的铺满山坡。一丝怅然恐慌成无法言说的失落。凝不住经年的妩媚,留不下往昔的绚烂。在季节的深处,九月的风微凉着心事。一场寒雨在风中呜咽叹息,轻敲轩窗,碎断人肠。没有了昨日纤雨湿花的雅致,忘却了曾经柳絮扑帘的悠然。绮梦悠悠伤情绵绵,一任愁绪在风中雨里飘摇成伤,恍若,枝头颤动的叶子,没了昨日张扬的碧绿,沉淀了一个春夏的水嫩在风中瞬间走失,只留下一袭苍白,挣扎在季节的末梢。嘴角溢出的笑,潋滟了一湖的秋水。如果出去嘴角的色斑,我想这个微笑会更美 脸上长色斑去除偏方, 《客户案例》   莫小姐   长时间面对电脑的坏处我一直都知道,像眼睛疲劳、干�� �,身体处于亚健康之类的, �� �电脑,好在我什么不舒服的症状都没出现,正当我庆幸不已� ��同时,我发现我额头、鼻梁两边和上眼睑处长斑了,数量很 多,就像雨后春笋一样,一点一点在短时间内全部冒出来了�� �看着脸上那如同苍蝇屎大小的斑点,颜色为黄色,略比肤色� ��点,远看近看都明显。哪个女孩子不爱美 为此特别烦恼。 b r   在网上搜索祛斑产品时发现了「黛芙薇尔精华液」,了�� �了一番之后觉得还不错,而且网上论坛里面大家发帖子都说� ��黛芙薇尔精华液」好,没发现有说副作用的,然后我还上「 黛芙薇尔精华液」的网站很认真的看了一下产品说明,说明�� �字很专业详细,对治疗机理、产品优势描述得挺不错的,我� ��在网上订购了一个周期的。   大概了十几天的样子,我鼻梁两边的斑块颜色就变淡了�� �面积也慢慢在缩小,本来粗糙、多油的皮肤状况也得到了改� ��。一个月过去以后,对着镜子看自己脸上的斑,鼻梁上的大 斑块已经缩小了一大半,颜色由以前的深褐色变成了浅褐色�� �额头和眼角处的斑也变淡变浅了。「黛芙薇尔精华液」祛斑� ��家在我订购产品的时候就说过,产品在祛斑的同时还能调经 ,让经期变得正常,还能减轻痛经的痛苦。我使用第二个周�� �产品的时候,这方面的效果显示出来了,我以前痛经非常厉� ��,自从使用这个套装以来,不但月经量变得正常了,而且痛 经也没那么严重了,在祛斑的同时还能调经,「黛芙薇尔精�� �液」真的很不错。 很快我就使用完第二个周期的产品,鼻梁两边的斑基本上看�� �到了,颜色跟肤色很接近,不仔细看根本看不出来了,我可� ��不用粉底盖住那一块的斑。额头上的斑也在慢慢消退,特别 是眼睑处的斑,淡化的比较明显。我接着使用第三个周期的�� �品,这个周期见效很快,刚半个月多一点,眼睑处的斑就消� ��了,褐色的斑融进皮肤里面,一点都看不出来了,额头上的 也差不多没有了。使用完三个周期的「黛芙薇尔精华液」之�� �,我从斑女人变成了一个水嫩白净的女人,而且我的月经也� ��得很有规律,内分泌正常,心情也变得非常开朗,哪位想要 祛斑的朋友就学我的祛斑方法,相信通过使用「黛芙薇尔精�� �液」,大家都能祛斑成功。 阅读了脸上长色斑去除偏方,再看脸上容易长斑的原因: 《色斑形成原因》   内部因素   一、压力   当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。   二、荷尔蒙分泌失调   避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加, — 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。   三、新陈代谢缓慢   肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。   四、错误的使用化妆品   使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。   外部因素   一、紫外线   照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。   二、不良的清洁习惯   因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。   三、遗传基因   父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》    黛芙薇尔精华液真的有效果吗 真的可以把脸上的黄褐�� �去掉吗   答:黛芙薇尔精华液dna精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客, 的新�� �客都是通过老顾客介绍而来,口碑由此而来    ,服用黛芙薇尔美白,会伤身体吗 有副作用吗   答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“dna美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作, �� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖    ,去除黄褐斑之后,会反弹吗   答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌 我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗    ,你们的价格有点贵,能不能便宜一点   答: , , ,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助 一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗 你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗    ,我适合用黛芙薇尔精华液吗   答:黛芙薇尔适用人群:    、生理紊乱引起的黄褐斑人群    、生育引起的妊娠斑人群    、年纪增长引起的老年斑人群    、化妆品色素沉积、辐射斑人群    、长期日照引起的日晒斑人群    、肌肤暗淡急需美白的人群 《祛斑小方法》 脸上长色斑去除偏方,同时为您分享祛斑小方法 一本中医古书上看到了一则去除雀斑的方子:将黄豆(黄豆�� �生的)浸泡在醋中一个月,每天服用几粒,坚持一段时间即� ��完全消除。 去斑讲究内外兼修,这点值得一试 。 original issue reported on code google com by additive gmail com on jul at
1
29,212
5,611,884,353
IssuesEvent
2017-04-03 01:21:31
prettydiff/prettydiff
https://api.github.com/repos/prettydiff/prettydiff
opened
New simplified diff algorithm broke diffcli output
Defect Diff Not started
Currently the diffcli option is over-engineered to compensate for a poor diff algorithm. The formatting and building of output needs to be an array of output prepared in the diffview.js library so that consuming interfaces simply loop through this array output with a console.log statement and without any additional logic.
1.0
New simplified diff algorithm broke diffcli output - Currently the diffcli option is over-engineered to compensate for a poor diff algorithm. The formatting and building of output needs to be an array of output prepared in the diffview.js library so that consuming interfaces simply loop through this array output with a console.log statement and without any additional logic.
defect
new simplified diff algorithm broke diffcli output currently the diffcli option is over engineered to compensate for a poor diff algorithm the formatting and building of output needs to be an array of output prepared in the diffview js library so that consuming interfaces simply loop through this array output with a console log statement and without any additional logic
1
29,928
5,959,397,176
IssuesEvent
2017-05-29 10:55:26
bridgedotnet/Bridge
https://api.github.com/repos/bridgedotnet/Bridge
closed
Enum.TryParse fails with exception
defect in progress
A description of the issue. ### Steps To Reproduce https://deck.net/08c324d3c077c47c7dade0e01405f1ec https://dotnetfiddle.net/Runi2c ```c# public enum Mode { None, A, B, C } public class Program { private static void Main(string[] args) { const string section = null; Mode mode; if (Enum.TryParse(section, true, out mode) == false) mode = Mode.A; Console.WriteLine(mode); } } ``` ### Expected Result ```js A ``` ### Actual Result ```js TypeError: Cannot read property 'toLowerCase' of null ```
1.0
Enum.TryParse fails with exception - A description of the issue. ### Steps To Reproduce https://deck.net/08c324d3c077c47c7dade0e01405f1ec https://dotnetfiddle.net/Runi2c ```c# public enum Mode { None, A, B, C } public class Program { private static void Main(string[] args) { const string section = null; Mode mode; if (Enum.TryParse(section, true, out mode) == false) mode = Mode.A; Console.WriteLine(mode); } } ``` ### Expected Result ```js A ``` ### Actual Result ```js TypeError: Cannot read property 'toLowerCase' of null ```
defect
enum tryparse fails with exception a description of the issue steps to reproduce c public enum mode none a b c public class program private static void main string args const string section null mode mode if enum tryparse section true out mode false mode mode a console writeline mode expected result js a actual result js typeerror cannot read property tolowercase of null
1
35,606
7,787,962,456
IssuesEvent
2018-06-07 01:35:55
google/sanitizers
https://api.github.com/repos/google/sanitizers
closed
Fine-grained origins for class members
Priority-Medium ProjectMemorySanitizer Status-Accepted Type-Defect
Originally reported on Google Code with ID 31 ``` Currently MSan reports the same origin for all members of dynamically allocated objects (pointing to the new() call for that object). It would be nice to include the actual member name in MSan report. Looks like we need Clang support for this - llvm IR does not have enough info (there are class layouts, but no member names). ``` Reported by `eugenis@google.com` on 2013-09-24 09:13:26
1.0
Fine-grained origins for class members - Originally reported on Google Code with ID 31 ``` Currently MSan reports the same origin for all members of dynamically allocated objects (pointing to the new() call for that object). It would be nice to include the actual member name in MSan report. Looks like we need Clang support for this - llvm IR does not have enough info (there are class layouts, but no member names). ``` Reported by `eugenis@google.com` on 2013-09-24 09:13:26
defect
fine grained origins for class members originally reported on google code with id currently msan reports the same origin for all members of dynamically allocated objects pointing to the new call for that object it would be nice to include the actual member name in msan report looks like we need clang support for this llvm ir does not have enough info there are class layouts but no member names reported by eugenis google com on
1
10,458
2,622,161,945
IssuesEvent
2015-03-04 00:10:25
byzhang/terrastore
https://api.github.com/repos/byzhang/terrastore
closed
Backup implementation for Ensembles
auto-migrated Milestone-0.8.2 Priority-High Project-Terrastore Type-Defect
``` How do you scale a currently running ensemble from N clusters to N+1 cluster without losing data? I had a ensemble running of 2 clusters (each with 1 master and 1 server) I tried the following, but could not retain all the data (I had just 1 bucket): 1. Export backup of the bucket of both the servers. 2. Restart the servers using the update cluster.json files. So now I have a ensemble of 3 cluster (the third cluster is also 1 master and 1 server) 3. Import the backup on both the servers. 4. When I try to GET all data, it shows 404 for some of the records. I even tried cleaning the masters before joining the servers in the new ensemble, but faced the same issue. ``` Original issue reported on code.google.com by `rohi...@gmail.com` on 10 Jun 2011 at 2:58
1.0
Backup implementation for Ensembles - ``` How do you scale a currently running ensemble from N clusters to N+1 cluster without losing data? I had a ensemble running of 2 clusters (each with 1 master and 1 server) I tried the following, but could not retain all the data (I had just 1 bucket): 1. Export backup of the bucket of both the servers. 2. Restart the servers using the update cluster.json files. So now I have a ensemble of 3 cluster (the third cluster is also 1 master and 1 server) 3. Import the backup on both the servers. 4. When I try to GET all data, it shows 404 for some of the records. I even tried cleaning the masters before joining the servers in the new ensemble, but faced the same issue. ``` Original issue reported on code.google.com by `rohi...@gmail.com` on 10 Jun 2011 at 2:58
defect
backup implementation for ensembles how do you scale a currently running ensemble from n clusters to n cluster without losing data i had a ensemble running of clusters each with master and server i tried the following but could not retain all the data i had just bucket export backup of the bucket of both the servers restart the servers using the update cluster json files so now i have a ensemble of cluster the third cluster is also master and server import the backup on both the servers when i try to get all data it shows for some of the records i even tried cleaning the masters before joining the servers in the new ensemble but faced the same issue original issue reported on code google com by rohi gmail com on jun at
1
135,454
12,684,739,392
IssuesEvent
2020-06-19 23:57:58
Azure/azure-cli
https://api.github.com/repos/Azure/azure-cli
closed
It's always good to add actual example of each command.
Documentation OKR3.4 Candidate
It's always good to add actual example of each command. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 9abfaa54-0895-29f7-f3bb-ff604912b2ac * Version Independent ID: d1ff3478-8a86-59e4-62ef-449900132a67 * Content: [az account](https://docs.microsoft.com/en-us/cli/azure/account?view=azure-cli-latest#az-account-set) * Content Source: [src/azure-cli/azure/cli/command_modules/profile/_help.py](https://github.com/Azure/azure-cli/blob/dev/src/azure-cli/azure/cli/command_modules/profile/_help.py) * Service: **azure** * GitHub Login: @rloutlaw * Microsoft Alias: **routlaw**
1.0
It's always good to add actual example of each command. - It's always good to add actual example of each command. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 9abfaa54-0895-29f7-f3bb-ff604912b2ac * Version Independent ID: d1ff3478-8a86-59e4-62ef-449900132a67 * Content: [az account](https://docs.microsoft.com/en-us/cli/azure/account?view=azure-cli-latest#az-account-set) * Content Source: [src/azure-cli/azure/cli/command_modules/profile/_help.py](https://github.com/Azure/azure-cli/blob/dev/src/azure-cli/azure/cli/command_modules/profile/_help.py) * Service: **azure** * GitHub Login: @rloutlaw * Microsoft Alias: **routlaw**
non_defect
it s always good to add actual example of each command it s always good to add actual example of each command document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service azure github login rloutlaw microsoft alias routlaw
0
10,513
2,622,169,415
IssuesEvent
2015-03-04 00:13:49
byzhang/rapidjson
https://api.github.com/repos/byzhang/rapidjson
closed
Unable to build tests on Mac OS X
auto-migrated Priority-Medium Type-Defect
``` I am trying to build the tests for rapidjson 0.11 (http://code.google.com/p/rapidjson/) on Mac OS X . It includes three projects: gtest (builds fine), unittest (build fails), and perftest (build fails), and when building make error out with Error 1 and Error 2. Actually, the first time I ran make, gtest and unittest built fine, but perftest failed. Now, unittest is failing as well. The compiler output shows the following errors for both unittest and perftest which causes make to fail: ../../include/rapidjson/reader.h: In function ‘const char* rapidjson::SkipWhitespace_SIMD(const char*)’: ../../include/rapidjson/reader.h:116: error: ‘_SIDD_UBYTE_OPS’ was not declared in this scope ../../include/rapidjson/reader.h:116: error: ‘_SIDD_CMP_EQUAL_ANY’ was not declared in this scope ../../include/rapidjson/reader.h:116: error: ‘_SIDD_BIT_MASK’ was not declared in this scope ../../include/rapidjson/reader.h:116: error: ‘_SIDD_NEGATIVE_POLARITY’ was not declared in this scope These pre-processor constants are related to SSE4 instructions. The makefile includes the -msse4.2 compiler switch to enable SSE4.2 support, and looking through the header files reveal that on OS X, both SSE4_1 and SSE4_2 pre-processor constants need to be defined for the SIDD... constants to be defined. For some reason, these SIDD... constants aren't being defined. Further research showed that the -msse4 switch enables support for both SSE4.1 and SSE4.2, so I tried changing the switch to -msse4, but it still errors out. Not sure if the -msse4.2 switch automatically defines SSE4_2__ , but I tried manually defining it, and sill no luck. NOTE: If you want to try building it yourself on Mac, you will need to download a different premake script file, as the included one doesn't work. You can download the corrected script from the attachment on the second post here https://code.google.com/p/rapidjson/issues/detail?id=54 Any ideas on how to get it building successfully on OS X ? ``` Original issue reported on code.google.com by `keith3...@rocketmail.com` on 23 Jan 2013 at 5:28 * Merged into: #55
1.0
Unable to build tests on Mac OS X - ``` I am trying to build the tests for rapidjson 0.11 (http://code.google.com/p/rapidjson/) on Mac OS X . It includes three projects: gtest (builds fine), unittest (build fails), and perftest (build fails), and when building make error out with Error 1 and Error 2. Actually, the first time I ran make, gtest and unittest built fine, but perftest failed. Now, unittest is failing as well. The compiler output shows the following errors for both unittest and perftest which causes make to fail: ../../include/rapidjson/reader.h: In function ‘const char* rapidjson::SkipWhitespace_SIMD(const char*)’: ../../include/rapidjson/reader.h:116: error: ‘_SIDD_UBYTE_OPS’ was not declared in this scope ../../include/rapidjson/reader.h:116: error: ‘_SIDD_CMP_EQUAL_ANY’ was not declared in this scope ../../include/rapidjson/reader.h:116: error: ‘_SIDD_BIT_MASK’ was not declared in this scope ../../include/rapidjson/reader.h:116: error: ‘_SIDD_NEGATIVE_POLARITY’ was not declared in this scope These pre-processor constants are related to SSE4 instructions. The makefile includes the -msse4.2 compiler switch to enable SSE4.2 support, and looking through the header files reveal that on OS X, both SSE4_1 and SSE4_2 pre-processor constants need to be defined for the SIDD... constants to be defined. For some reason, these SIDD... constants aren't being defined. Further research showed that the -msse4 switch enables support for both SSE4.1 and SSE4.2, so I tried changing the switch to -msse4, but it still errors out. Not sure if the -msse4.2 switch automatically defines SSE4_2__ , but I tried manually defining it, and sill no luck. NOTE: If you want to try building it yourself on Mac, you will need to download a different premake script file, as the included one doesn't work. You can download the corrected script from the attachment on the second post here https://code.google.com/p/rapidjson/issues/detail?id=54 Any ideas on how to get it building successfully on OS X ? ``` Original issue reported on code.google.com by `keith3...@rocketmail.com` on 23 Jan 2013 at 5:28 * Merged into: #55
defect
unable to build tests on mac os x i am trying to build the tests for rapidjson on mac os x it includes three projects gtest builds fine unittest build fails and perftest build fails and when building make error out with error and error actually the first time i ran make gtest and unittest built fine but perftest failed now unittest is failing as well the compiler output shows the following errors for both unittest and perftest which causes make to fail include rapidjson reader h in function ‘const char rapidjson skipwhitespace simd const char ’ include rapidjson reader h error ‘ sidd ubyte ops’ was not declared in this scope include rapidjson reader h error ‘ sidd cmp equal any’ was not declared in this scope include rapidjson reader h error ‘ sidd bit mask’ was not declared in this scope include rapidjson reader h error ‘ sidd negative polarity’ was not declared in this scope these pre processor constants are related to instructions the makefile includes the compiler switch to enable support and looking through the header files reveal that on os x both and pre processor constants need to be defined for the sidd constants to be defined for some reason these sidd constants aren t being defined further research showed that the switch enables support for both and so i tried changing the switch to but it still errors out not sure if the switch automatically defines but i tried manually defining it and sill no luck note if you want to try building it yourself on mac you will need to download a different premake script file as the included one doesn t work you can download the corrected script from the attachment on the second post here any ideas on how to get it building successfully on os x original issue reported on code google com by rocketmail com on jan at merged into
1
46,033
13,055,841,879
IssuesEvent
2020-07-30 02:53:52
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
opened
icerec/trunk dox on software.icecube.wisc.edu are "403 Forbidden" (Trac #474)
Incomplete Migration Migrated from Trac combo reconstruction defect
Migrated from https://code.icecube.wisc.edu/ticket/474 ```json { "status": "closed", "changetime": "2014-10-03T18:27:09", "description": "On http://software.icecube.wisc.edu/ there is a link to IceRec \"nightly builds\": http://software.icecube.wisc.edu/icerec_trunk/ which results in \"403 - Forbidden\".\n\nI am assigning this now to dladieu, but I could equally well imagine that this is actually the job of the icerec metaproject coordinator, which would be Meike de With these days.", "reporter": "boersma", "cc": "meike.dewith", "resolution": "fixed", "_ts": "1412360829211490", "component": "combo reconstruction", "summary": "icerec/trunk dox on software.icecube.wisc.edu are \"403 Forbidden\"", "priority": "minor", "keywords": "icerec documentation doxygen", "time": "2013-11-26T10:42:01", "milestone": "", "owner": "nega", "type": "defect" } ```
1.0
icerec/trunk dox on software.icecube.wisc.edu are "403 Forbidden" (Trac #474) - Migrated from https://code.icecube.wisc.edu/ticket/474 ```json { "status": "closed", "changetime": "2014-10-03T18:27:09", "description": "On http://software.icecube.wisc.edu/ there is a link to IceRec \"nightly builds\": http://software.icecube.wisc.edu/icerec_trunk/ which results in \"403 - Forbidden\".\n\nI am assigning this now to dladieu, but I could equally well imagine that this is actually the job of the icerec metaproject coordinator, which would be Meike de With these days.", "reporter": "boersma", "cc": "meike.dewith", "resolution": "fixed", "_ts": "1412360829211490", "component": "combo reconstruction", "summary": "icerec/trunk dox on software.icecube.wisc.edu are \"403 Forbidden\"", "priority": "minor", "keywords": "icerec documentation doxygen", "time": "2013-11-26T10:42:01", "milestone": "", "owner": "nega", "type": "defect" } ```
defect
icerec trunk dox on software icecube wisc edu are forbidden trac migrated from json status closed changetime description on there is a link to icerec nightly builds which results in forbidden n ni am assigning this now to dladieu but i could equally well imagine that this is actually the job of the icerec metaproject coordinator which would be meike de with these days reporter boersma cc meike dewith resolution fixed ts component combo reconstruction summary icerec trunk dox on software icecube wisc edu are forbidden priority minor keywords icerec documentation doxygen time milestone owner nega type defect
1
36,275
7,875,182,247
IssuesEvent
2018-06-25 19:32:51
bridgedotnet/Bridge
https://api.github.com/repos/bridgedotnet/Bridge
closed
Inheriting from an external class (no base() call case)
defect
A class-successor has to call a default constructor of an external base class even when `base()` call is not specified explicitly. Related to #2735 and #2189. ### Steps To Reproduce https://deck.net/abf6fe87cb3641732d977c0f8543788b ```csharp public class Program { [Init(InitPosition.Top)] public static void Init() { /*@ var A = (function () { function A() { System.Console.WriteLine("A initialized."); } return A; }()); */ } public static void Main() { var b = new B(); } } [Virtual] [Namespace(false)] public class A { } public class B : A { public B() // : base() // Uncomment to fix { } } ``` ### Expected Result Console output: ``` A initialized. ``` ### Actual Result Nothing was written to output.
1.0
Inheriting from an external class (no base() call case) - A class-successor has to call a default constructor of an external base class even when `base()` call is not specified explicitly. Related to #2735 and #2189. ### Steps To Reproduce https://deck.net/abf6fe87cb3641732d977c0f8543788b ```csharp public class Program { [Init(InitPosition.Top)] public static void Init() { /*@ var A = (function () { function A() { System.Console.WriteLine("A initialized."); } return A; }()); */ } public static void Main() { var b = new B(); } } [Virtual] [Namespace(false)] public class A { } public class B : A { public B() // : base() // Uncomment to fix { } } ``` ### Expected Result Console output: ``` A initialized. ``` ### Actual Result Nothing was written to output.
defect
inheriting from an external class no base call case a class successor has to call a default constructor of an external base class even when base call is not specified explicitly related to and steps to reproduce csharp public class program public static void init var a function function a system console writeline a initialized return a public static void main var b new b public class a public class b a public b base uncomment to fix expected result console output a initialized actual result nothing was written to output
1
24,489
3,992,496,991
IssuesEvent
2016-05-10 02:06:50
cakephp/cakephp
https://api.github.com/repos/cakephp/cakephp
closed
missing 'all' in cake Completion subcommands bake
console Defect
This is a (multiple allowed): * [x] bug * [ ] enhancement * [ ] feature-discussion (RFC) * CakePHP Version: CakePHP v3.2.8 PHP : 7.0.6-1+donate.sury.org~trusty+1 ### What you did ```shell bin/cake Completion subcommands bake ``` ### Expected Behavior i'm expecting _all_ in command list ### Actual Behavior ``` [ bin/cake Completion subcommands bake master * ] 11:20 behavior cell component controller fixture form helper mailer migration migration_snapshot model plugin seed shell shell_helper task template test ```
1.0
missing 'all' in cake Completion subcommands bake - This is a (multiple allowed): * [x] bug * [ ] enhancement * [ ] feature-discussion (RFC) * CakePHP Version: CakePHP v3.2.8 PHP : 7.0.6-1+donate.sury.org~trusty+1 ### What you did ```shell bin/cake Completion subcommands bake ``` ### Expected Behavior i'm expecting _all_ in command list ### Actual Behavior ``` [ bin/cake Completion subcommands bake master * ] 11:20 behavior cell component controller fixture form helper mailer migration migration_snapshot model plugin seed shell shell_helper task template test ```
defect
missing all in cake completion subcommands bake this is a multiple allowed bug enhancement feature discussion rfc cakephp version cakephp php donate sury org trusty what you did shell bin cake completion subcommands bake expected behavior i m expecting all in command list actual behavior behavior cell component controller fixture form helper mailer migration migration snapshot model plugin seed shell shell helper task template test
1
72,441
24,119,803,869
IssuesEvent
2022-09-20 17:38:36
SeleniumHQ/selenium
https://api.github.com/repos/SeleniumHQ/selenium
opened
[🐛 Bug]: isPromise() returning false for thenable objects
I-defect needs-triaging
### What happened? https://github.com/SeleniumHQ/selenium/commit/84dd6109ce692788467432ccee55f481fe49f2bc#r84520555 In the above commit, the `isPromise()` utility function was simplified to look for `typeof value === '[object Promise]'`, however this fails the function's JSDoc documentation that any object exposing a `then()` method would be considered a Promise. My scripts that rely on `WebDriver.wait(<WebElementPromise>)` currently fail, but work after restoring the above change. I propose that the previous functionality is returned to the `isPromise()` function to allow thenable objects to be treated as Promises. ### How can we reproduce the issue? ```shell Pass an object with a `then()` method (for example a `WebElementPromise`) to `isPromise()`, it will return `false`. ``` ### Relevant log output ```shell n/a ``` ### Operating System Mac OS ### Selenium version selenium-webdriver@4.4.0 ### What are the browser(s) and version(s) where you see this issue? Chrome ### What are the browser driver(s) and version(s) where you see this issue? ChromeDriver 105.0.5195.19 (b9c217c128c16f53d12f9a02933fcfdec1bf49af-refs/branch-heads/5195@{#176}) ### Are you using Selenium Grid? _No response_
1.0
[🐛 Bug]: isPromise() returning false for thenable objects - ### What happened? https://github.com/SeleniumHQ/selenium/commit/84dd6109ce692788467432ccee55f481fe49f2bc#r84520555 In the above commit, the `isPromise()` utility function was simplified to look for `typeof value === '[object Promise]'`, however this fails the function's JSDoc documentation that any object exposing a `then()` method would be considered a Promise. My scripts that rely on `WebDriver.wait(<WebElementPromise>)` currently fail, but work after restoring the above change. I propose that the previous functionality is returned to the `isPromise()` function to allow thenable objects to be treated as Promises. ### How can we reproduce the issue? ```shell Pass an object with a `then()` method (for example a `WebElementPromise`) to `isPromise()`, it will return `false`. ``` ### Relevant log output ```shell n/a ``` ### Operating System Mac OS ### Selenium version selenium-webdriver@4.4.0 ### What are the browser(s) and version(s) where you see this issue? Chrome ### What are the browser driver(s) and version(s) where you see this issue? ChromeDriver 105.0.5195.19 (b9c217c128c16f53d12f9a02933fcfdec1bf49af-refs/branch-heads/5195@{#176}) ### Are you using Selenium Grid? _No response_
defect
ispromise returning false for thenable objects what happened in the above commit the ispromise utility function was simplified to look for typeof value however this fails the function s jsdoc documentation that any object exposing a then method would be considered a promise my scripts that rely on webdriver wait currently fail but work after restoring the above change i propose that the previous functionality is returned to the ispromise function to allow thenable objects to be treated as promises how can we reproduce the issue shell pass an object with a then method for example a webelementpromise to ispromise it will return false relevant log output shell n a operating system mac os selenium version selenium webdriver what are the browser s and version s where you see this issue chrome what are the browser driver s and version s where you see this issue chromedriver refs branch heads are you using selenium grid no response
1
11,654
2,660,023,444
IssuesEvent
2015-03-19 01:41:49
perfsonar/project
https://api.github.com/repos/perfsonar/project
closed
Review configuration examples in ls_registration_daemon.conf file
Component-LSRegistrationDaemon Milestone-Release3.4.2 Priority-Medium Type-Defect
Original [issue 1046](https://code.google.com/p/perfsonar-ps/issues/detail?id=1046) created by arlake228 on 2015-01-15T08:48:09.000Z: <b>What steps will reproduce the problem?</b> In the last suggestion from the mailing list it was written that the ls_registration_daemon.conf configuration file should have entries like: &lt;service_template&gt; &lt;local_latency_service&gt; autodiscover_addresses 1 primary_interface p1p1 &lt;/local_latency_service&gt; <b>What is the expected output? What do you see instead?</b> However looking at https://code.google.com/p/perfsonar-ps/source/browse/perfSONAR_PS-LSRegistrationDaemon/etc/ls_registration_daemon.conf?name=3.4 the suggested configuration syntax in comments is different: &lt;service_template local_latency_service&gt; autodiscover_addresses 1 # primary_address eth0 &lt;/service_template&gt; &lt;service_template local_bandwidth_service&gt; autodiscover_addresses 1 # primary_address eth1 &lt;/service_template&gt; It seems to be wrong and/or outdated. I suggest to review the file and correct to provide correct examples for users.
1.0
Review configuration examples in ls_registration_daemon.conf file - Original [issue 1046](https://code.google.com/p/perfsonar-ps/issues/detail?id=1046) created by arlake228 on 2015-01-15T08:48:09.000Z: <b>What steps will reproduce the problem?</b> In the last suggestion from the mailing list it was written that the ls_registration_daemon.conf configuration file should have entries like: &lt;service_template&gt; &lt;local_latency_service&gt; autodiscover_addresses 1 primary_interface p1p1 &lt;/local_latency_service&gt; <b>What is the expected output? What do you see instead?</b> However looking at https://code.google.com/p/perfsonar-ps/source/browse/perfSONAR_PS-LSRegistrationDaemon/etc/ls_registration_daemon.conf?name=3.4 the suggested configuration syntax in comments is different: &lt;service_template local_latency_service&gt; autodiscover_addresses 1 # primary_address eth0 &lt;/service_template&gt; &lt;service_template local_bandwidth_service&gt; autodiscover_addresses 1 # primary_address eth1 &lt;/service_template&gt; It seems to be wrong and/or outdated. I suggest to review the file and correct to provide correct examples for users.
defect
review configuration examples in ls registration daemon conf file original created by on what steps will reproduce the problem in the last suggestion from the mailing list it was written that the ls registration daemon conf configuration file should have entries like lt service template gt lt local latency service gt autodiscover addresses primary interface lt local latency service gt what is the expected output what do you see instead however looking at the suggested configuration syntax in comments is different lt service template local latency service gt autodiscover addresses primary address lt service template gt lt service template local bandwidth service gt autodiscover addresses primary address lt service template gt it seems to be wrong and or outdated i suggest to review the file and correct to provide correct examples for users
1
290,201
25,042,595,410
IssuesEvent
2022-11-04 23:01:30
godotengine/godot
https://api.github.com/repos/godotengine/godot
closed
Undo Close Tab is unreliable
bug topic:editor usability needs testing
<!-- Please search existing issues for potential duplicates before filing yours: https://github.com/godotengine/godot/issues?q=is%3Aissue --> **Godot version:** <!-- Specify commit hash if using non-official build. --> 3.2.2 **Issue description:** <!-- What happened, and what was expected. --> Unfortunately I don't have exact reproduction steps, but sometimes when you close scene tab and use "Undo Close Tab" (in my case with shortcut), a different tab is opened. Previously I thought that the order might get messed up, but now I think that some scenes just don't get saved in scene history and trying to reopen them reopens a different tab. This seems to happen randomly and then can be reproduced multiple times on that scene, but eventually fixes itself. **Steps to reproduce:** 1. Open a scene 2. Open another scene 3. Close both scenes 4. Reopen scene tab 5. If you are unlucky, wrong scene gets reopened
1.0
Undo Close Tab is unreliable - <!-- Please search existing issues for potential duplicates before filing yours: https://github.com/godotengine/godot/issues?q=is%3Aissue --> **Godot version:** <!-- Specify commit hash if using non-official build. --> 3.2.2 **Issue description:** <!-- What happened, and what was expected. --> Unfortunately I don't have exact reproduction steps, but sometimes when you close scene tab and use "Undo Close Tab" (in my case with shortcut), a different tab is opened. Previously I thought that the order might get messed up, but now I think that some scenes just don't get saved in scene history and trying to reopen them reopens a different tab. This seems to happen randomly and then can be reproduced multiple times on that scene, but eventually fixes itself. **Steps to reproduce:** 1. Open a scene 2. Open another scene 3. Close both scenes 4. Reopen scene tab 5. If you are unlucky, wrong scene gets reopened
non_defect
undo close tab is unreliable please search existing issues for potential duplicates before filing yours godot version issue description unfortunately i don t have exact reproduction steps but sometimes when you close scene tab and use undo close tab in my case with shortcut a different tab is opened previously i thought that the order might get messed up but now i think that some scenes just don t get saved in scene history and trying to reopen them reopens a different tab this seems to happen randomly and then can be reproduced multiple times on that scene but eventually fixes itself steps to reproduce open a scene open another scene close both scenes reopen scene tab if you are unlucky wrong scene gets reopened
0
228
2,520,557,016
IssuesEvent
2015-01-19 05:07:10
AtlasOfLivingAustralia/biocache-hubs
https://api.github.com/repos/AtlasOfLivingAustralia/biocache-hubs
closed
Fish or Fishes
priority-medium status-new type-defect
*migrated from:* https://code.google.com/p/ala/issues/detail?id=98 *date:* Thu Aug 8 16:27:28 2013 *author:* moyesyside --- Original Issue reported by Reported by john.t...@austmus.gov.au, Today (16 hours ago) - [https://code.google.com/p/ala-portal/issues/detail?id=300](https://code.google.com/p/ala-portal/issues/detail?id=300) Reported by john.tann@austmus.gov.au, Today (16 hours ago) I have been told that when we refer to 'fish' as a lifeform we should use the word 'fishes' See [http://australianmuseum.net.au/fish-and-fishes](http://australianmuseum.net.au/fish-and-fishes) The word 'fish' should be changed to 'fishes' when referring to a lifeform. **What steps will reproduce the problem?** An example: Australian Museum Ichthyology collection. in the graph of records by higher level group: Axis is Arthropods Fish Insects Reptiles **What is the expected output? What do you see instead?** Axis is Arthropods ***Fishes*** Insects Reptiles What version of the product are you using? On what operating system? 2013 Please provide any additional information below. The term 'fish' is probably incorrectly used in a bunch of places. This is a low priority fix. fish or fishes.png 4.7 KB   View   Download
1.0
Fish or Fishes - *migrated from:* https://code.google.com/p/ala/issues/detail?id=98 *date:* Thu Aug 8 16:27:28 2013 *author:* moyesyside --- Original Issue reported by Reported by john.t...@austmus.gov.au, Today (16 hours ago) - [https://code.google.com/p/ala-portal/issues/detail?id=300](https://code.google.com/p/ala-portal/issues/detail?id=300) Reported by john.tann@austmus.gov.au, Today (16 hours ago) I have been told that when we refer to 'fish' as a lifeform we should use the word 'fishes' See [http://australianmuseum.net.au/fish-and-fishes](http://australianmuseum.net.au/fish-and-fishes) The word 'fish' should be changed to 'fishes' when referring to a lifeform. **What steps will reproduce the problem?** An example: Australian Museum Ichthyology collection. in the graph of records by higher level group: Axis is Arthropods Fish Insects Reptiles **What is the expected output? What do you see instead?** Axis is Arthropods ***Fishes*** Insects Reptiles What version of the product are you using? On what operating system? 2013 Please provide any additional information below. The term 'fish' is probably incorrectly used in a bunch of places. This is a low priority fix. fish or fishes.png 4.7 KB   View   Download
defect
fish or fishes migrated from date thu aug author moyesyside original issue reported by reported by john t austmus gov au today hours ago reported by john tann austmus gov au today hours ago i have been told that when we refer to fish as a lifeform we should use the word fishes see the word fish should be changed to fishes when referring to a lifeform what steps will reproduce the problem an example australian museum ichthyology collection in the graph of records by higher level group axis is arthropods fish insects reptiles what is the expected output what do you see instead axis is arthropods fishes insects reptiles what version of the product are you using on what operating system please provide any additional information below the term fish is probably incorrectly used in a bunch of places this is a low priority fix fish or fishes png kb   view   download
1
83,031
10,318,815,523
IssuesEvent
2019-08-30 15:49:04
CICE-Consortium/CICE
https://api.github.com/repos/CICE-Consortium/CICE
closed
evp kernel version 2 testing and validation
Documentation Dynamics Priority: High Testing Type: Feature
We are going to merge PR #278, PR #252. There are several outstanding issues, basically copied from the end of #252, --------------------- Let me summarize where we are. With evp_kernel_ver=0, results are bit-for-bit for most tests against the current master. This is running full test suites on gordon for 4 compilers. A subset of box tests are NOT bit-for-bit on 3/4 compilers. Rerunning the failed box tests with the debug flag (reduced optimization and run time checks) on both master and this PR results in bit-for-bit identical answers. It seems the changes in the answers in the box test is caused by some compiler optimization as a results of the code changes. This might be associated with the evp kernel changes (although @mhrib makes a case it shouldn't) or it might be associated with some of the code cleanup. We could look into this further or we could accept it. Personally, I am comfortable with this outcome as it stands. I believe we've shown the answers are roundoff different (see above gbox128 diff) as a result of compiler optimization and that we can make this bit-for-bit if we reduce compiler optimization. I think based on these results, we could merge this PR. evp_kernel_ver=0 will be the default setting. Separately, there is an effort to test and validate the evp_kernel_ver=2. The same test suite on gordon was run with the new kernel on. Results can be found https://github.com/CICE-Consortium/Test-Results/wiki/cice_by_hash_forks, hash aa6de33...+evpk=2. Three to four tests fail on each compiler, and they are the same tests across the compilers. Looking at the intel results, https://github.com/CICE-Consortium/Test-Results/wiki/aa6de33f19.gordon.pgi.190128.235649, there are four failures. - restart gbox128 4x2. This test runs but fails to restart exactly. - restart gx1 40x4 droundrobin medium. This test fails with "(abort_ice) error = (horizontal_remap)ERROR: bad departure points" on the first timestep. - restart gx3 16x2x5x10x20 drakeX2. This test fails with "(abort_ice) error = (horizontal_remap)ERROR: bad departure points" on the first timestep. - restart tx1 40x4 dsectrobin medium. This test fails gracefully in the evp kernel. tx1 is not supported yet. Again, many tests passed, but these 4 failures need to be debugged. In addition, the qc test relies on the gx1 configuration, so the qc testing comparing evp_kernel_ver=2 to 0 could not be done. So, the outstanding tasks are - debug the 4 failures noted above - run the qc test comparing evp_kernel_ver=0 to evp_kernel_ver=2. This requires gx1 (one of the failing tests) - update documentation - change evp_kernel_ver variable to kevp_kernel - produce and document timing information comparing evp_kernel_ver=0 and 2. - add evp_kernel_ver=2 tests to the test suite - maybe do a little cleanup on ice_dyn_evp_1d.F90 to make the code a little more readable (breaks between subroutines and such)
1.0
evp kernel version 2 testing and validation - We are going to merge PR #278, PR #252. There are several outstanding issues, basically copied from the end of #252, --------------------- Let me summarize where we are. With evp_kernel_ver=0, results are bit-for-bit for most tests against the current master. This is running full test suites on gordon for 4 compilers. A subset of box tests are NOT bit-for-bit on 3/4 compilers. Rerunning the failed box tests with the debug flag (reduced optimization and run time checks) on both master and this PR results in bit-for-bit identical answers. It seems the changes in the answers in the box test is caused by some compiler optimization as a results of the code changes. This might be associated with the evp kernel changes (although @mhrib makes a case it shouldn't) or it might be associated with some of the code cleanup. We could look into this further or we could accept it. Personally, I am comfortable with this outcome as it stands. I believe we've shown the answers are roundoff different (see above gbox128 diff) as a result of compiler optimization and that we can make this bit-for-bit if we reduce compiler optimization. I think based on these results, we could merge this PR. evp_kernel_ver=0 will be the default setting. Separately, there is an effort to test and validate the evp_kernel_ver=2. The same test suite on gordon was run with the new kernel on. Results can be found https://github.com/CICE-Consortium/Test-Results/wiki/cice_by_hash_forks, hash aa6de33...+evpk=2. Three to four tests fail on each compiler, and they are the same tests across the compilers. Looking at the intel results, https://github.com/CICE-Consortium/Test-Results/wiki/aa6de33f19.gordon.pgi.190128.235649, there are four failures. - restart gbox128 4x2. This test runs but fails to restart exactly. - restart gx1 40x4 droundrobin medium. This test fails with "(abort_ice) error = (horizontal_remap)ERROR: bad departure points" on the first timestep. - restart gx3 16x2x5x10x20 drakeX2. This test fails with "(abort_ice) error = (horizontal_remap)ERROR: bad departure points" on the first timestep. - restart tx1 40x4 dsectrobin medium. This test fails gracefully in the evp kernel. tx1 is not supported yet. Again, many tests passed, but these 4 failures need to be debugged. In addition, the qc test relies on the gx1 configuration, so the qc testing comparing evp_kernel_ver=2 to 0 could not be done. So, the outstanding tasks are - debug the 4 failures noted above - run the qc test comparing evp_kernel_ver=0 to evp_kernel_ver=2. This requires gx1 (one of the failing tests) - update documentation - change evp_kernel_ver variable to kevp_kernel - produce and document timing information comparing evp_kernel_ver=0 and 2. - add evp_kernel_ver=2 tests to the test suite - maybe do a little cleanup on ice_dyn_evp_1d.F90 to make the code a little more readable (breaks between subroutines and such)
non_defect
evp kernel version testing and validation we are going to merge pr pr there are several outstanding issues basically copied from the end of let me summarize where we are with evp kernel ver results are bit for bit for most tests against the current master this is running full test suites on gordon for compilers a subset of box tests are not bit for bit on compilers rerunning the failed box tests with the debug flag reduced optimization and run time checks on both master and this pr results in bit for bit identical answers it seems the changes in the answers in the box test is caused by some compiler optimization as a results of the code changes this might be associated with the evp kernel changes although mhrib makes a case it shouldn t or it might be associated with some of the code cleanup we could look into this further or we could accept it personally i am comfortable with this outcome as it stands i believe we ve shown the answers are roundoff different see above diff as a result of compiler optimization and that we can make this bit for bit if we reduce compiler optimization i think based on these results we could merge this pr evp kernel ver will be the default setting separately there is an effort to test and validate the evp kernel ver the same test suite on gordon was run with the new kernel on results can be found hash evpk three to four tests fail on each compiler and they are the same tests across the compilers looking at the intel results there are four failures restart this test runs but fails to restart exactly restart droundrobin medium this test fails with abort ice error horizontal remap error bad departure points on the first timestep restart this test fails with abort ice error horizontal remap error bad departure points on the first timestep restart dsectrobin medium this test fails gracefully in the evp kernel is not supported yet again many tests passed but these failures need to be debugged in addition the qc test relies on the configuration so the qc testing comparing evp kernel ver to could not be done so the outstanding tasks are debug the failures noted above run the qc test comparing evp kernel ver to evp kernel ver this requires one of the failing tests update documentation change evp kernel ver variable to kevp kernel produce and document timing information comparing evp kernel ver and add evp kernel ver tests to the test suite maybe do a little cleanup on ice dyn evp to make the code a little more readable breaks between subroutines and such
0
27,811
5,106,884,223
IssuesEvent
2017-01-05 13:10:36
TASVideos/BizHawk
https://api.github.com/repos/TASVideos/BizHawk
closed
Genplus-gx [BizHawk] Game Genie codes not working
bug Core-EmuHawk Core-Genplus-GX Core-GensHawk OpSys-Any Priority-Medium Type-Defect
I can't get any Game Genie codes to work on BizHawk Genplus-gx. I have codes that I know that work but on BizHawk they don't work. I know how to use BizHawk. Like on the game Strider (UE) [!]. Here is a code that give you Infinite Life In The Life Gauge. AK8T-AA5R AKVT-AA94 AJLT-AA9E On Kega this code works.
1.0
Genplus-gx [BizHawk] Game Genie codes not working - I can't get any Game Genie codes to work on BizHawk Genplus-gx. I have codes that I know that work but on BizHawk they don't work. I know how to use BizHawk. Like on the game Strider (UE) [!]. Here is a code that give you Infinite Life In The Life Gauge. AK8T-AA5R AKVT-AA94 AJLT-AA9E On Kega this code works.
defect
genplus gx game genie codes not working i can t get any game genie codes to work on bizhawk genplus gx i have codes that i know that work but on bizhawk they don t work i know how to use bizhawk like on the game strider ue here is a code that give you infinite life in the life gauge akvt ajlt on kega this code works
1
277,173
21,016,821,004
IssuesEvent
2022-03-30 11:49:32
SW-Team2/se-team2-tetris
https://api.github.com/repos/SW-Team2/se-team2-tetris
closed
docs: PR, Issue template 제작
📄 documentation
## 개요 - PR과 Issue을 보면서 명확하게 어떤 것을 했는지 전달이 안되는 부분이 있었습니다. - 예를 들어, 어떤 기능을 구현했을 때, 이것이 어떻게 동작하는지 캡쳐 화면이라던가 동영상 등 추가적인 정보가 없어 직접 실행해야 하는 수고로움이 있었습니다. - 탬플릿을 제공하여 다른 사람에게 명확하게 정보를 제공하고자 합니다. ## 할 일 - [x] PR template 제작 - [x] Issue template 제작 ## 참고 https://docs.github.com/en/communities/using-templates-to-encourage-useful-issues-and-pull-requests/about-issue-and-pull-request-templates
1.0
docs: PR, Issue template 제작 - ## 개요 - PR과 Issue을 보면서 명확하게 어떤 것을 했는지 전달이 안되는 부분이 있었습니다. - 예를 들어, 어떤 기능을 구현했을 때, 이것이 어떻게 동작하는지 캡쳐 화면이라던가 동영상 등 추가적인 정보가 없어 직접 실행해야 하는 수고로움이 있었습니다. - 탬플릿을 제공하여 다른 사람에게 명확하게 정보를 제공하고자 합니다. ## 할 일 - [x] PR template 제작 - [x] Issue template 제작 ## 참고 https://docs.github.com/en/communities/using-templates-to-encourage-useful-issues-and-pull-requests/about-issue-and-pull-request-templates
non_defect
docs pr issue template 제작 개요 pr과 issue을 보면서 명확하게 어떤 것을 했는지 전달이 안되는 부분이 있었습니다 예를 들어 어떤 기능을 구현했을 때 이것이 어떻게 동작하는지 캡쳐 화면이라던가 동영상 등 추가적인 정보가 없어 직접 실행해야 하는 수고로움이 있었습니다 탬플릿을 제공하여 다른 사람에게 명확하게 정보를 제공하고자 합니다 할 일 pr template 제작 issue template 제작 참고
0
225,023
17,789,594,198
IssuesEvent
2021-08-31 14:48:40
elastic/kibana
https://api.github.com/repos/elastic/kibana
opened
[RAC] [Observability] Create functional tests for alerts table add-to-case actions
test Team:logs-metrics-ui Theme: rac
## :notebook: Summary We should create functional tests for the add-to-case actions on the Observability Alerts page. :link: makes use of the service created in #110627 ## :heavy_check_mark: Test cases :warning: When you pick this up, please spend a minute to think about and expand on these test cases. This is just an initial proposal. - add-to-case row actions are... - enabled with `all` permissions for the different observability solutions owning the respective rules - enabled for APM alerts with special APM alert permissions - disabled with `read` and `none` permissions for the different observability solutions owning the respective rules
1.0
[RAC] [Observability] Create functional tests for alerts table add-to-case actions - ## :notebook: Summary We should create functional tests for the add-to-case actions on the Observability Alerts page. :link: makes use of the service created in #110627 ## :heavy_check_mark: Test cases :warning: When you pick this up, please spend a minute to think about and expand on these test cases. This is just an initial proposal. - add-to-case row actions are... - enabled with `all` permissions for the different observability solutions owning the respective rules - enabled for APM alerts with special APM alert permissions - disabled with `read` and `none` permissions for the different observability solutions owning the respective rules
non_defect
create functional tests for alerts table add to case actions notebook summary we should create functional tests for the add to case actions on the observability alerts page link makes use of the service created in heavy check mark test cases warning when you pick this up please spend a minute to think about and expand on these test cases this is just an initial proposal add to case row actions are enabled with all permissions for the different observability solutions owning the respective rules enabled for apm alerts with special apm alert permissions disabled with read and none permissions for the different observability solutions owning the respective rules
0
46,225
13,055,872,553
IssuesEvent
2020-07-30 02:59:16
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
opened
Matplotlib crashes on systems with AMD bulldozer architecture. (Trac #795)
Incomplete Migration Migrated from Trac defect tools/ports
Migrated from https://code.icecube.wisc.edu/ticket/795 ```json { "status": "closed", "changetime": "2014-10-30T18:25:07", "description": "Matplotlib provided by cvmfs crashes on SL6 with AMD bulldozer processors (AMD FX-8350).\n{{{\npython -c \"import pylab; pylab.plot(1,2); pylab.show()\"\n}}}\n\n\ncrashes with the message\n{{{ \nIllegal instruction\n}}}\n\nGDB hints at libgoto as the source of the problem (see attachment)\n", "reporter": "chaack", "cc": "nega", "resolution": "fixed", "_ts": "1414693507102825", "component": "tools/ports", "summary": "Matplotlib crashes on systems with AMD bulldozer architecture.", "priority": "major", "keywords": "BLAS cvmfs", "time": "2014-10-25T14:34:25", "milestone": "", "owner": "david.schultz", "type": "defect" } ```
1.0
Matplotlib crashes on systems with AMD bulldozer architecture. (Trac #795) - Migrated from https://code.icecube.wisc.edu/ticket/795 ```json { "status": "closed", "changetime": "2014-10-30T18:25:07", "description": "Matplotlib provided by cvmfs crashes on SL6 with AMD bulldozer processors (AMD FX-8350).\n{{{\npython -c \"import pylab; pylab.plot(1,2); pylab.show()\"\n}}}\n\n\ncrashes with the message\n{{{ \nIllegal instruction\n}}}\n\nGDB hints at libgoto as the source of the problem (see attachment)\n", "reporter": "chaack", "cc": "nega", "resolution": "fixed", "_ts": "1414693507102825", "component": "tools/ports", "summary": "Matplotlib crashes on systems with AMD bulldozer architecture.", "priority": "major", "keywords": "BLAS cvmfs", "time": "2014-10-25T14:34:25", "milestone": "", "owner": "david.schultz", "type": "defect" } ```
defect
matplotlib crashes on systems with amd bulldozer architecture trac migrated from json status closed changetime description matplotlib provided by cvmfs crashes on with amd bulldozer processors amd fx n npython c import pylab pylab plot pylab show n n n ncrashes with the message n nillegal instruction n n ngdb hints at libgoto as the source of the problem see attachment n reporter chaack cc nega resolution fixed ts component tools ports summary matplotlib crashes on systems with amd bulldozer architecture priority major keywords blas cvmfs time milestone owner david schultz type defect
1
9,750
2,615,167,204
IssuesEvent
2015-03-01 06:48:01
chrsmith/reaver-wps
https://api.github.com/repos/chrsmith/reaver-wps
opened
99.99% Reaver repeats same pin witn no m5 packets
auto-migrated Priority-Triage Type-Defect
``` 0. What version of Reaver are you using? (Only defects against the latest version will be considered.) , reaver 1.4 1. What operating system are you using (Linux is the only supported OS)? Backtrack 5 r3 usb boot 2. Is your wireless card in monitor mode (yes/no)? yes 3. What is the signal strength of the Access Point you are trying to crack? -60 4. What is the manufacturer and model # of the device you are trying to crack? BtHub4 5. What is the entire command line string you are supplying to reaver? reaver -i mon0 -b CC:BB:33:A3:BD:7C -d 0 -t 0 -N -S 6. Please describe what you think the issue is. Hi everyone. For some reason I run an attack against an AP and I got to 99.99% with pin tstarting with 998 well i mean what are the chances that the password is not in the range of 0 to 998 .... and no m5 packets are sent. or recieved therefore I assume the first part of the pin was incorrect... If so then reaver missed the correct pin? I recieved Warning: Timeout and 0.02 and 0.03 error, however the cracking progressed... Also I noticed that when crakcking reaver does this: Trying pin 23456789 Trying pin 23457780 Trying pin 23458781 Something like this, therefore Im not sure but is it possible that reaver missed the correct pin?? 7. Paste the output from Reaver below. Sadly I do not have it. I'm writing from Windows and I failed to record the log however m1 packets m2 packet m3 packet m4 packet Wsc Nack Eapol REquest Trying pin XXXXXXX ^repeat.... and then it goes to 99.99% and it took me 20 hours to do it.. repeats the same pin over and over which start with 998.......... chances are so low...... ``` Original issue reported on code.google.com by `Krer...@gmail.com` on 12 Nov 2013 at 12:15
1.0
99.99% Reaver repeats same pin witn no m5 packets - ``` 0. What version of Reaver are you using? (Only defects against the latest version will be considered.) , reaver 1.4 1. What operating system are you using (Linux is the only supported OS)? Backtrack 5 r3 usb boot 2. Is your wireless card in monitor mode (yes/no)? yes 3. What is the signal strength of the Access Point you are trying to crack? -60 4. What is the manufacturer and model # of the device you are trying to crack? BtHub4 5. What is the entire command line string you are supplying to reaver? reaver -i mon0 -b CC:BB:33:A3:BD:7C -d 0 -t 0 -N -S 6. Please describe what you think the issue is. Hi everyone. For some reason I run an attack against an AP and I got to 99.99% with pin tstarting with 998 well i mean what are the chances that the password is not in the range of 0 to 998 .... and no m5 packets are sent. or recieved therefore I assume the first part of the pin was incorrect... If so then reaver missed the correct pin? I recieved Warning: Timeout and 0.02 and 0.03 error, however the cracking progressed... Also I noticed that when crakcking reaver does this: Trying pin 23456789 Trying pin 23457780 Trying pin 23458781 Something like this, therefore Im not sure but is it possible that reaver missed the correct pin?? 7. Paste the output from Reaver below. Sadly I do not have it. I'm writing from Windows and I failed to record the log however m1 packets m2 packet m3 packet m4 packet Wsc Nack Eapol REquest Trying pin XXXXXXX ^repeat.... and then it goes to 99.99% and it took me 20 hours to do it.. repeats the same pin over and over which start with 998.......... chances are so low...... ``` Original issue reported on code.google.com by `Krer...@gmail.com` on 12 Nov 2013 at 12:15
defect
reaver repeats same pin witn no packets what version of reaver are you using only defects against the latest version will be considered reaver what operating system are you using linux is the only supported os backtrack usb boot is your wireless card in monitor mode yes no yes what is the signal strength of the access point you are trying to crack what is the manufacturer and model of the device you are trying to crack what is the entire command line string you are supplying to reaver reaver i b cc bb bd d t n s please describe what you think the issue is hi everyone for some reason i run an attack against an ap and i got to with pin tstarting with well i mean what are the chances that the password is not in the range of to and no packets are sent or recieved therefore i assume the first part of the pin was incorrect if so then reaver missed the correct pin i recieved warning timeout and and error however the cracking progressed also i noticed that when crakcking reaver does this trying pin trying pin trying pin something like this therefore im not sure but is it possible that reaver missed the correct pin paste the output from reaver below sadly i do not have it i m writing from windows and i failed to record the log however packets packet packet packet wsc nack eapol request trying pin xxxxxxx repeat and then it goes to and it took me hours to do it repeats the same pin over and over which start with chances are so low original issue reported on code google com by krer gmail com on nov at
1
440,647
30,754,420,055
IssuesEvent
2023-07-28 23:34:16
aws-samples/amazon-kinesis-video-streams-demos
https://api.github.com/repos/aws-samples/amazon-kinesis-video-streams-demos
closed
Pass AWS and stream link arguments in docker run
documentation question
How can I pass the: `AWS_ACCESS_KEY_ID=<AWS_ACCESS_KEY_ID> AWS_SECRET_ACCESS_KEY=<AWS_SECRET_ACCESS_KEY> ./kvs_gstreamer_sample <STREAM_NAME> <RTSP_URL> ` Directly in the docker run?
1.0
Pass AWS and stream link arguments in docker run - How can I pass the: `AWS_ACCESS_KEY_ID=<AWS_ACCESS_KEY_ID> AWS_SECRET_ACCESS_KEY=<AWS_SECRET_ACCESS_KEY> ./kvs_gstreamer_sample <STREAM_NAME> <RTSP_URL> ` Directly in the docker run?
non_defect
pass aws and stream link arguments in docker run how can i pass the aws access key id aws secret access key kvs gstreamer sample directly in the docker run
0
23,541
4,021,604,118
IssuesEvent
2016-05-16 22:40:30
elastic/logstash
https://api.github.com/repos/elastic/logstash
closed
Hot Threads failure in Travis CI
bug tests
Using our new [Travis CI branch](https://github.com/elastic/logstash/pull/4608) we sometimes get [this odd failure](https://travis-ci.org/elastic/logstash/builds/124009233). This seems like an intermittently failing test we'll have to look into. @purbon do you understand what's going on here?
1.0
Hot Threads failure in Travis CI - Using our new [Travis CI branch](https://github.com/elastic/logstash/pull/4608) we sometimes get [this odd failure](https://travis-ci.org/elastic/logstash/builds/124009233). This seems like an intermittently failing test we'll have to look into. @purbon do you understand what's going on here?
non_defect
hot threads failure in travis ci using our new we sometimes get this seems like an intermittently failing test we ll have to look into purbon do you understand what s going on here
0
43,241
23,163,299,355
IssuesEvent
2022-07-29 20:28:01
mattermost/focalboard
https://api.github.com/repos/mattermost/focalboard
closed
PERF: Fetch board members in parallel
Bug Sev/1 Performance
Here in web app codebase, we fetch each individual board member's details in a separate API call - https://github.com/mattermost/focalboard/blob/3c7fd72dcf0456e14b324cfb0621371289603f2a/webapp/src/store/boards.ts#L36 I propose fetching users in bulk instead of one by one.
True
PERF: Fetch board members in parallel - Here in web app codebase, we fetch each individual board member's details in a separate API call - https://github.com/mattermost/focalboard/blob/3c7fd72dcf0456e14b324cfb0621371289603f2a/webapp/src/store/boards.ts#L36 I propose fetching users in bulk instead of one by one.
non_defect
perf fetch board members in parallel here in web app codebase we fetch each individual board member s details in a separate api call i propose fetching users in bulk instead of one by one
0
489,843
14,112,715,084
IssuesEvent
2020-11-07 07:04:53
AY2021S1-CS2103T-W12-4/tp
https://api.github.com/repos/AY2021S1-CS2103T-W12-4/tp
closed
Update UG - Consistent examples across UG and code
priority.High type.Task type.UserGuide
TODOS: 1. Update command summary section 2. Change parameter name for students from `NAME` to `STUDENT_NAME` 3. For individual commands, update to the latest examples (ref google docs) 4. For commands with more than 1 method, label each of the formats to the respective method (ie `markpresent`)
1.0
Update UG - Consistent examples across UG and code - TODOS: 1. Update command summary section 2. Change parameter name for students from `NAME` to `STUDENT_NAME` 3. For individual commands, update to the latest examples (ref google docs) 4. For commands with more than 1 method, label each of the formats to the respective method (ie `markpresent`)
non_defect
update ug consistent examples across ug and code todos update command summary section change parameter name for students from name to student name for individual commands update to the latest examples ref google docs for commands with more than method label each of the formats to the respective method ie markpresent
0
361,307
10,707,104,460
IssuesEvent
2019-10-24 16:43:11
netdata/netdata
https://api.github.com/repos/netdata/netdata
closed
Generate alarms if the disk cannot keep up with data collection
area/database feature request priority/high
<!--- When creating a feature request please: - Verify first that your issue is not already reported on GitHub - Explain new feature briefly in "Feature idea summary" section - Provide a clear and concise description of what you expect to happen. ---> ##### Feature idea summary If the disk cannot keep up with metric collection this condition should be detectable and alarms should be generated. This is for `memory mode = dbengine`.
1.0
Generate alarms if the disk cannot keep up with data collection - <!--- When creating a feature request please: - Verify first that your issue is not already reported on GitHub - Explain new feature briefly in "Feature idea summary" section - Provide a clear and concise description of what you expect to happen. ---> ##### Feature idea summary If the disk cannot keep up with metric collection this condition should be detectable and alarms should be generated. This is for `memory mode = dbengine`.
non_defect
generate alarms if the disk cannot keep up with data collection when creating a feature request please verify first that your issue is not already reported on github explain new feature briefly in feature idea summary section provide a clear and concise description of what you expect to happen feature idea summary if the disk cannot keep up with metric collection this condition should be detectable and alarms should be generated this is for memory mode dbengine
0
46,383
5,806,359,928
IssuesEvent
2017-05-04 02:19:00
NSW-OEH-EMS-KST/grid-garage-3
https://api.github.com/repos/NSW-OEH-EMS-KST/grid-garage-3
closed
raster - tweak values
tested and working
Let users know that the min and max values must be integers.. @byezy this still the case?
1.0
raster - tweak values - Let users know that the min and max values must be integers.. @byezy this still the case?
non_defect
raster tweak values let users know that the min and max values must be integers byezy this still the case
0
70,673
23,282,203,022
IssuesEvent
2022-08-05 13:11:52
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
Error configuring hazelcast with DSL style "IMap eviction config doesn't support max size policy `ENTRY_COUNT`"
Type: Defect
<!-- Thanks for reporting your issue. Please share with us the following information, to help us resolve your issue quickly and efficiently. --> **Describe the bug** When i configure a eviction in a map (please see shared code) , once the map is getting used the app throws error (see exception below) **Expected behavior** I expect to get the map **To Reproduce** Steps to reproduce the behavior: 1. Setup a map with the eviction LRU or LFU or NONE. 2. Use by the map 3. Error is thrown 4. The app is in clustered but for this case is just a single instance. **Additional context** <!-- Add any other context about the problem here. Common details that we're often interested in: - Detailed description of the steps to reproduce your issue - Logs and stack traces, if available - Hazelcast version that you use (e.g. 3.4, also specify whether it is a minor release or the latest snapshot) - If available, integration module versions (e.g. Tomcat, Jetty, Spring, Hibernate). Also, include their detailed configuration information such as web.xml, Hibernate configuration and `context.xml` for Spring - Cluster size, i.e. the number of Hazelcast cluster members - Number of the clients - Version of Java. It is also helpful to mention the JVM parameters - Operating system. If it is Linux, kernel version is helpful - Unit test with the `hazelcast.xml` file. If you could include a unit test which reproduces your issue, we would be grateful --> -The application is a spring boot app. -I'm using spring-hazelcast dependency that aims to use 5.1 version of hazelcast core. -I tried updating to 5.1.2 , same behavior. -Spring boot version is 2.6.8 Config config = new XmlConfigBuilder().build(); config.setClusterName("cluster-name"); config .setInstanceName(instanceName) .addMapConfig(new MapConfig() .setBackupCount(1) .setName("example-map") .setEvictionConfig(new EvictionConfig().setEvictionPolicy(LFU)) .setMapStoreConfig(new MapStoreConfig() .setInitialLoadMode(EAGER) .setWriteBatchSize(500) .setWriteCoalescing(true) .setEnabled(true) .setFactoryImplementation(mapStoreConfig) .setWriteDelaySeconds(5))); Logs: `com.hazelcast.config.InvalidConfigurationException: IMap eviction config doesn't support max size policy `ENTRY_COUNT`. Please select a valid one: [PER_NODE, PER_PARTITION, USED_HEAP_PERCENTAGE, USED_HEAP_SIZE, FREE_HEAP_PERCENTAGE, FREE_HEAP_SIZE, USED_NATIVE_MEMORY_SIZE, USED_NATIVE_MEMORY_PERCENTAGE, FREE_NATIVE_MEMORY_SIZE, FREE_NATIVE_MEMORY_PERCENTAGE]. at com.hazelcast.internal.config.ConfigValidator.checkMapMaxSizePolicyConfig(ConfigValidator.java:199) at com.hazelcast.internal.config.ConfigValidator.checkMapEvictionConfig(ConfigValidator.java:177)`
1.0
Error configuring hazelcast with DSL style "IMap eviction config doesn't support max size policy `ENTRY_COUNT`" - <!-- Thanks for reporting your issue. Please share with us the following information, to help us resolve your issue quickly and efficiently. --> **Describe the bug** When i configure a eviction in a map (please see shared code) , once the map is getting used the app throws error (see exception below) **Expected behavior** I expect to get the map **To Reproduce** Steps to reproduce the behavior: 1. Setup a map with the eviction LRU or LFU or NONE. 2. Use by the map 3. Error is thrown 4. The app is in clustered but for this case is just a single instance. **Additional context** <!-- Add any other context about the problem here. Common details that we're often interested in: - Detailed description of the steps to reproduce your issue - Logs and stack traces, if available - Hazelcast version that you use (e.g. 3.4, also specify whether it is a minor release or the latest snapshot) - If available, integration module versions (e.g. Tomcat, Jetty, Spring, Hibernate). Also, include their detailed configuration information such as web.xml, Hibernate configuration and `context.xml` for Spring - Cluster size, i.e. the number of Hazelcast cluster members - Number of the clients - Version of Java. It is also helpful to mention the JVM parameters - Operating system. If it is Linux, kernel version is helpful - Unit test with the `hazelcast.xml` file. If you could include a unit test which reproduces your issue, we would be grateful --> -The application is a spring boot app. -I'm using spring-hazelcast dependency that aims to use 5.1 version of hazelcast core. -I tried updating to 5.1.2 , same behavior. -Spring boot version is 2.6.8 Config config = new XmlConfigBuilder().build(); config.setClusterName("cluster-name"); config .setInstanceName(instanceName) .addMapConfig(new MapConfig() .setBackupCount(1) .setName("example-map") .setEvictionConfig(new EvictionConfig().setEvictionPolicy(LFU)) .setMapStoreConfig(new MapStoreConfig() .setInitialLoadMode(EAGER) .setWriteBatchSize(500) .setWriteCoalescing(true) .setEnabled(true) .setFactoryImplementation(mapStoreConfig) .setWriteDelaySeconds(5))); Logs: `com.hazelcast.config.InvalidConfigurationException: IMap eviction config doesn't support max size policy `ENTRY_COUNT`. Please select a valid one: [PER_NODE, PER_PARTITION, USED_HEAP_PERCENTAGE, USED_HEAP_SIZE, FREE_HEAP_PERCENTAGE, FREE_HEAP_SIZE, USED_NATIVE_MEMORY_SIZE, USED_NATIVE_MEMORY_PERCENTAGE, FREE_NATIVE_MEMORY_SIZE, FREE_NATIVE_MEMORY_PERCENTAGE]. at com.hazelcast.internal.config.ConfigValidator.checkMapMaxSizePolicyConfig(ConfigValidator.java:199) at com.hazelcast.internal.config.ConfigValidator.checkMapEvictionConfig(ConfigValidator.java:177)`
defect
error configuring hazelcast with dsl style imap eviction config doesn t support max size policy entry count thanks for reporting your issue please share with us the following information to help us resolve your issue quickly and efficiently describe the bug when i configure a eviction in a map please see shared code once the map is getting used the app throws error see exception below expected behavior i expect to get the map to reproduce steps to reproduce the behavior setup a map with the eviction lru or lfu or none use by the map error is thrown the app is in clustered but for this case is just a single instance additional context add any other context about the problem here common details that we re often interested in detailed description of the steps to reproduce your issue logs and stack traces if available hazelcast version that you use e g also specify whether it is a minor release or the latest snapshot if available integration module versions e g tomcat jetty spring hibernate also include their detailed configuration information such as web xml hibernate configuration and context xml for spring cluster size i e the number of hazelcast cluster members number of the clients version of java it is also helpful to mention the jvm parameters operating system if it is linux kernel version is helpful unit test with the hazelcast xml file if you could include a unit test which reproduces your issue we would be grateful the application is a spring boot app i m using spring hazelcast dependency that aims to use version of hazelcast core i tried updating to same behavior spring boot version is config config new xmlconfigbuilder build config setclustername cluster name config setinstancename instancename addmapconfig new mapconfig setbackupcount setname example map setevictionconfig new evictionconfig setevictionpolicy lfu setmapstoreconfig new mapstoreconfig setinitialloadmode eager setwritebatchsize setwritecoalescing true setenabled true setfactoryimplementation mapstoreconfig setwritedelayseconds logs com hazelcast config invalidconfigurationexception imap eviction config doesn t support max size policy entry count please select a valid one at com hazelcast internal config configvalidator checkmapmaxsizepolicyconfig configvalidator java at com hazelcast internal config configvalidator checkmapevictionconfig configvalidator java
1
27,094
4,875,252,256
IssuesEvent
2016-11-16 08:57:56
TNGSB/eWallet
https://api.github.com/repos/TNGSB/eWallet
closed
eWallet_MobileApp(Airtime)_Both IOS & Android #097
Defect - Medium (Sev-3)
[Defect_Mobile App #97.xlsx](https://github.com/TNGSB/eWallet/files/591389/Defect_Mobile.App.97.xlsx) Test Description : To verify the error message when user left all the fields in blank Defect Description : System displayed wrong error message when user left all the fields in blank - apply to both IOS and Android Refer attachment for POT
1.0
eWallet_MobileApp(Airtime)_Both IOS & Android #097 - [Defect_Mobile App #97.xlsx](https://github.com/TNGSB/eWallet/files/591389/Defect_Mobile.App.97.xlsx) Test Description : To verify the error message when user left all the fields in blank Defect Description : System displayed wrong error message when user left all the fields in blank - apply to both IOS and Android Refer attachment for POT
defect
ewallet mobileapp airtime both ios android test description to verify the error message when user left all the fields in blank defect description system displayed wrong error message when user left all the fields in blank apply to both ios and android refer attachment for pot
1
594,799
18,054,428,399
IssuesEvent
2021-09-20 05:48:22
naev/naev
https://api.github.com/repos/naev/naev
closed
mission markers are hard to distinguish
Type-Enhancement Priority-Low
I find it hard to distinguish the active mission markers from the markers for potential missions that I could accept. I understand you got rid of colours there for colourblind accessibility, but what about making them different shapes, or use a different phase for the blinking? I don't mean frequency, too high might trigger epilepsy and too low makes it hard to discern (there is still some usable range there, but that'd probably better be used for mission priorities), but different phase at same frequency would work for proposed mission maker - of which there is only one - vs. active markers - of which there are usually multiple. If the phase difference is noticably different from pi, it will even be possible to tell which is which when both are in view with a bit of getting used to it. I think something like PI/2 or 2/3 PI should do it (where time between two maxima of the same marker is 2*PI). Logically, I'd think the phase of markers for missions already taken should lead the phase of markers for missions that could potentially be taken.
1.0
mission markers are hard to distinguish - I find it hard to distinguish the active mission markers from the markers for potential missions that I could accept. I understand you got rid of colours there for colourblind accessibility, but what about making them different shapes, or use a different phase for the blinking? I don't mean frequency, too high might trigger epilepsy and too low makes it hard to discern (there is still some usable range there, but that'd probably better be used for mission priorities), but different phase at same frequency would work for proposed mission maker - of which there is only one - vs. active markers - of which there are usually multiple. If the phase difference is noticably different from pi, it will even be possible to tell which is which when both are in view with a bit of getting used to it. I think something like PI/2 or 2/3 PI should do it (where time between two maxima of the same marker is 2*PI). Logically, I'd think the phase of markers for missions already taken should lead the phase of markers for missions that could potentially be taken.
non_defect
mission markers are hard to distinguish i find it hard to distinguish the active mission markers from the markers for potential missions that i could accept i understand you got rid of colours there for colourblind accessibility but what about making them different shapes or use a different phase for the blinking i don t mean frequency too high might trigger epilepsy and too low makes it hard to discern there is still some usable range there but that d probably better be used for mission priorities but different phase at same frequency would work for proposed mission maker of which there is only one vs active markers of which there are usually multiple if the phase difference is noticably different from pi it will even be possible to tell which is which when both are in view with a bit of getting used to it i think something like pi or pi should do it where time between two maxima of the same marker is pi logically i d think the phase of markers for missions already taken should lead the phase of markers for missions that could potentially be taken
0
331,522
28,967,621,427
IssuesEvent
2023-05-10 08:59:38
Joystream/joystream
https://api.github.com/repos/Joystream/joystream
opened
Fix direct channel payment flow
network-integration-test nara-network
Error is introduced in merging `nara` -> `crt_release` output from processor: ![image](https://github.com/Joystream/joystream/assets/51839739/3076aa21-c9e2-4f4d-8d0f-c9a8bd549ea2) failing flow: https://github.com/Joystream/joystream/actions/runs/4926868090/jobs/8817633132?pr=4749
1.0
Fix direct channel payment flow - Error is introduced in merging `nara` -> `crt_release` output from processor: ![image](https://github.com/Joystream/joystream/assets/51839739/3076aa21-c9e2-4f4d-8d0f-c9a8bd549ea2) failing flow: https://github.com/Joystream/joystream/actions/runs/4926868090/jobs/8817633132?pr=4749
non_defect
fix direct channel payment flow error is introduced in merging nara crt release output from processor failing flow
0
50,900
13,187,954,763
IssuesEvent
2020-08-13 05:07:50
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
closed
[filterscripts] coordinate service -> astro (Trac #1641)
Migrated from Trac combo reconstruction defect
This import needs to change: ```text File "/home/dschultz/Documents/combo/trunk/build_memory/lib/icecube/filterscripts/gcfilter.py", line 17, in GCFilter from icecube import dataclasses, coordinate_service ImportError: cannot import name coordinate_service ``` And this line in the file: ```text add_gcfilter(tray,label=name,If=If,CorsikaMJDStart=coordinate_service.calendar_date_2_mjd(2013,07,01,00,00,00.),CorsikaMJDEnd=coordinate_service.calendar_date_2_mjd(2013,07,01,23,56,04.091)) ##One full MJD ``` <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1641">https://code.icecube.wisc.edu/ticket/1641</a>, reported by david.schultz and owned by kjmeagher</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:11:57", "description": "This import needs to change:\n{{{\n File \"/home/dschultz/Documents/combo/trunk/build_memory/lib/icecube/filterscripts/gcfilter.py\", line 17, in GCFilter\n from icecube import dataclasses, coordinate_service\nImportError: cannot import name coordinate_service\n}}}\n\nAnd this line in the file:\n{{{\nadd_gcfilter(tray,label=name,If=If,CorsikaMJDStart=coordinate_service.calendar_date_2_mjd(2013,07,01,00,00,00.),CorsikaMJDEnd=coordinate_service.calendar_date_2_mjd(2013,07,01,23,56,04.091)) ##One full MJD\n}}}", "reporter": "david.schultz", "cc": "", "resolution": "fixed", "_ts": "1550067117911749", "component": "combo reconstruction", "summary": "[filterscripts] coordinate service -> astro", "priority": "blocker", "keywords": "", "time": "2016-04-17T17:21:55", "milestone": "", "owner": "kjmeagher", "type": "defect" } ``` </p> </details>
1.0
[filterscripts] coordinate service -> astro (Trac #1641) - This import needs to change: ```text File "/home/dschultz/Documents/combo/trunk/build_memory/lib/icecube/filterscripts/gcfilter.py", line 17, in GCFilter from icecube import dataclasses, coordinate_service ImportError: cannot import name coordinate_service ``` And this line in the file: ```text add_gcfilter(tray,label=name,If=If,CorsikaMJDStart=coordinate_service.calendar_date_2_mjd(2013,07,01,00,00,00.),CorsikaMJDEnd=coordinate_service.calendar_date_2_mjd(2013,07,01,23,56,04.091)) ##One full MJD ``` <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1641">https://code.icecube.wisc.edu/ticket/1641</a>, reported by david.schultz and owned by kjmeagher</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:11:57", "description": "This import needs to change:\n{{{\n File \"/home/dschultz/Documents/combo/trunk/build_memory/lib/icecube/filterscripts/gcfilter.py\", line 17, in GCFilter\n from icecube import dataclasses, coordinate_service\nImportError: cannot import name coordinate_service\n}}}\n\nAnd this line in the file:\n{{{\nadd_gcfilter(tray,label=name,If=If,CorsikaMJDStart=coordinate_service.calendar_date_2_mjd(2013,07,01,00,00,00.),CorsikaMJDEnd=coordinate_service.calendar_date_2_mjd(2013,07,01,23,56,04.091)) ##One full MJD\n}}}", "reporter": "david.schultz", "cc": "", "resolution": "fixed", "_ts": "1550067117911749", "component": "combo reconstruction", "summary": "[filterscripts] coordinate service -> astro", "priority": "blocker", "keywords": "", "time": "2016-04-17T17:21:55", "milestone": "", "owner": "kjmeagher", "type": "defect" } ``` </p> </details>
defect
coordinate service astro trac this import needs to change text file home dschultz documents combo trunk build memory lib icecube filterscripts gcfilter py line in gcfilter from icecube import dataclasses coordinate service importerror cannot import name coordinate service and this line in the file text add gcfilter tray label name if if corsikamjdstart coordinate service calendar date mjd corsikamjdend coordinate service calendar date mjd one full mjd migrated from json status closed changetime description this import needs to change n n file home dschultz documents combo trunk build memory lib icecube filterscripts gcfilter py line in gcfilter n from icecube import dataclasses coordinate service nimporterror cannot import name coordinate service n n nand this line in the file n nadd gcfilter tray label name if if corsikamjdstart coordinate service calendar date mjd corsikamjdend coordinate service calendar date mjd one full mjd n reporter david schultz cc resolution fixed ts component combo reconstruction summary coordinate service astro priority blocker keywords time milestone owner kjmeagher type defect
1
122,655
10,228,902,302
IssuesEvent
2019-08-17 07:28:33
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
closed
JDBC driver installation related info missing in JDBC api doc
Area/StandardLibs BetaTesting Type/Docs
**Description:** Subject please in doc https://v1-0-0-alpha.ballerina.io/learn/api-docs/ballerina/java.jdbc/index.html **Steps to reproduce:** **Affected Versions:** **OS, DB, other environment details and versions:** **Related Issues (optional):** <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --> **Suggested Labels (optional):** <!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels--> **Suggested Assignees (optional):** <!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
1.0
JDBC driver installation related info missing in JDBC api doc - **Description:** Subject please in doc https://v1-0-0-alpha.ballerina.io/learn/api-docs/ballerina/java.jdbc/index.html **Steps to reproduce:** **Affected Versions:** **OS, DB, other environment details and versions:** **Related Issues (optional):** <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --> **Suggested Labels (optional):** <!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels--> **Suggested Assignees (optional):** <!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
non_defect
jdbc driver installation related info missing in jdbc api doc description subject please in doc steps to reproduce affected versions os db other environment details and versions related issues optional suggested labels optional suggested assignees optional
0
426,942
29,669,335,257
IssuesEvent
2023-06-11 07:51:16
fedewf1/repositorio-tp2
https://api.github.com/repos/fedewf1/repositorio-tp2
closed
Issues 2 tp5
documentation Diseño
**Para los archivos sucursales.html, contacto.html, sus derivados archivos html y los archivos java que empleen. Aplicar lo siguiente. **No tener en cuenta los archivos nav, header y footer. Dado que estas son comunues a todos los archivos. Considerar el uso de las clases bootstrap Containers, Filas, Columnas, Alineación, Formularios y componentes que considere necesario. El sitio se debe adaptar (ser responsive) para los diferentes tamaños de dispositivos. Adaptar las clases del paquete model para anotarlas con @Component. En las clases del paquete controller realizar la inyeccion de dependencia de los objetos necesarios para el controlador. Agregar documentacion interna al codigo de los archivos mencionados al principio. Dentro del paquete model realice la agregación de las anotaciones necesarias para realizar la validación de los atributos. En los controladores que procesan las peticiones de formularios realice las adaptaciones necesarias para capturar los errores. Realizar las modificaciones dentro de los archivos html que representan formularios para poder mostrar al usuario los mensajes referidos a las validaciones.
1.0
Issues 2 tp5 - **Para los archivos sucursales.html, contacto.html, sus derivados archivos html y los archivos java que empleen. Aplicar lo siguiente. **No tener en cuenta los archivos nav, header y footer. Dado que estas son comunues a todos los archivos. Considerar el uso de las clases bootstrap Containers, Filas, Columnas, Alineación, Formularios y componentes que considere necesario. El sitio se debe adaptar (ser responsive) para los diferentes tamaños de dispositivos. Adaptar las clases del paquete model para anotarlas con @Component. En las clases del paquete controller realizar la inyeccion de dependencia de los objetos necesarios para el controlador. Agregar documentacion interna al codigo de los archivos mencionados al principio. Dentro del paquete model realice la agregación de las anotaciones necesarias para realizar la validación de los atributos. En los controladores que procesan las peticiones de formularios realice las adaptaciones necesarias para capturar los errores. Realizar las modificaciones dentro de los archivos html que representan formularios para poder mostrar al usuario los mensajes referidos a las validaciones.
non_defect
issues para los archivos sucursales html contacto html sus derivados archivos html y los archivos java que empleen aplicar lo siguiente no tener en cuenta los archivos nav header y footer dado que estas son comunues a todos los archivos considerar el uso de las clases bootstrap containers filas columnas alineación formularios y componentes que considere necesario el sitio se debe adaptar ser responsive para los diferentes tamaños de dispositivos adaptar las clases del paquete model para anotarlas con component en las clases del paquete controller realizar la inyeccion de dependencia de los objetos necesarios para el controlador agregar documentacion interna al codigo de los archivos mencionados al principio dentro del paquete model realice la agregación de las anotaciones necesarias para realizar la validación de los atributos en los controladores que procesan las peticiones de formularios realice las adaptaciones necesarias para capturar los errores realizar las modificaciones dentro de los archivos html que representan formularios para poder mostrar al usuario los mensajes referidos a las validaciones
0
52,093
13,211,387,926
IssuesEvent
2020-08-15 22:46:50
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
opened
[iceprod2] handle expiration of self-signed cert for webserver (Trac #1676)
Incomplete Migration Migrated from Trac defect iceprod
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1676">https://code.icecube.wisc.edu/projects/icecube/ticket/1676</a>, reported by david.schultzand owned by david.schultz</em></summary> <p> ```json { "status": "closed", "changetime": "2016-05-09T21:55:16", "_ts": "1462830916934929", "description": "The default right now is to generate a self-signed cert with a duration of 30 days. We could either make this time infinitely large (several years), or deal with the problem of expiring certs.\n\nNote that just making a new cert isn't good enough, because already running jobs will still have the old cert.", "reporter": "david.schultz", "cc": "", "resolution": "wontfix", "time": "2016-04-29T16:41:30", "component": "iceprod", "summary": "[iceprod2] handle expiration of self-signed cert for webserver", "priority": "major", "keywords": "", "milestone": "", "owner": "david.schultz", "type": "defect" } ``` </p> </details>
1.0
[iceprod2] handle expiration of self-signed cert for webserver (Trac #1676) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1676">https://code.icecube.wisc.edu/projects/icecube/ticket/1676</a>, reported by david.schultzand owned by david.schultz</em></summary> <p> ```json { "status": "closed", "changetime": "2016-05-09T21:55:16", "_ts": "1462830916934929", "description": "The default right now is to generate a self-signed cert with a duration of 30 days. We could either make this time infinitely large (several years), or deal with the problem of expiring certs.\n\nNote that just making a new cert isn't good enough, because already running jobs will still have the old cert.", "reporter": "david.schultz", "cc": "", "resolution": "wontfix", "time": "2016-04-29T16:41:30", "component": "iceprod", "summary": "[iceprod2] handle expiration of self-signed cert for webserver", "priority": "major", "keywords": "", "milestone": "", "owner": "david.schultz", "type": "defect" } ``` </p> </details>
defect
handle expiration of self signed cert for webserver trac migrated from json status closed changetime ts description the default right now is to generate a self signed cert with a duration of days we could either make this time infinitely large several years or deal with the problem of expiring certs n nnote that just making a new cert isn t good enough because already running jobs will still have the old cert reporter david schultz cc resolution wontfix time component iceprod summary handle expiration of self signed cert for webserver priority major keywords milestone owner david schultz type defect
1
30,869
6,335,280,954
IssuesEvent
2017-07-26 18:30:13
scipy/scipy
https://api.github.com/repos/scipy/scipy
closed
optimize.BenchGlobal broken
Benchmarks defect
As reported on the mailing list, ``optimize.BenchGlobal`` benchmarks are broken currently: ``` python runtests.py --bench optimize.BenchGlobal --> fails on both my modified version (to add stochasticBB testing) and on the original scipy repository (errors described in the pastebin above) python runtests.py --bench optimize.BenchLeastSquares --> works flawlessly python runtests.py --bench optimize.BenchSmoothUnbounded --> works flawlessly ``` Traceback: =============== =============== ==================== ============= ================== -- result type / solver --------------- --------------------------------------------------------------------- test function success% / DE success% / basinh. <nfev> / DE <nfev> / basinh. =============== =============== ==================== ============= ================== AMGM failed failed failed failed =============== =============== ==================== ============= ================== [100.00%] ····· For parameters: 'AMGM', 'success%', 'DE' Traceback (most recent call last): File "/Library/Frameworks/Python.framework/Versions/3.5/lib/python3.5/site-packages/asv/benchmark.py", line 773, in <module> commands[mode](args) File "/Library/Frameworks/Python.framework/Versions/3.5/lib/python3.5/site-packages/asv/benchmark.py", line 744, in main_run skip = benchmark.do_setup() File "/Library/Frameworks/Python.framework/Versions/3.5/lib/python3.5/site-packages/asv/benchmark.py", line 325, in do_setup setup(*self._current_params) TypeError: setup() takes 4 positional arguments but 5 were given ... Version information: Python 3.5 numpy 1.13.0 scipy 1.0.0.dev0+32d8e9e (current master) asv 0.3.dev1123+32d09e76 (current master) The problem seems to be that an extra parameter is passed from somewhere: ``setup() takes 4 positional arguments but 5 were given``. Checking those params shows that what's passed is ``name, ret_value, solver, None``. Unclear where the ``None`` is produced exactly. The ``BenchGlobal`` constructor which does not call ``super(BenchGlobal, self).__init__`` seems fragile. Changing the signature of setup by appending a ``none`` parameter fixes things, but that doesn't look quite right. Last time this was modified is in gh-6400. @andyfaff what do you think?
1.0
optimize.BenchGlobal broken - As reported on the mailing list, ``optimize.BenchGlobal`` benchmarks are broken currently: ``` python runtests.py --bench optimize.BenchGlobal --> fails on both my modified version (to add stochasticBB testing) and on the original scipy repository (errors described in the pastebin above) python runtests.py --bench optimize.BenchLeastSquares --> works flawlessly python runtests.py --bench optimize.BenchSmoothUnbounded --> works flawlessly ``` Traceback: =============== =============== ==================== ============= ================== -- result type / solver --------------- --------------------------------------------------------------------- test function success% / DE success% / basinh. <nfev> / DE <nfev> / basinh. =============== =============== ==================== ============= ================== AMGM failed failed failed failed =============== =============== ==================== ============= ================== [100.00%] ····· For parameters: 'AMGM', 'success%', 'DE' Traceback (most recent call last): File "/Library/Frameworks/Python.framework/Versions/3.5/lib/python3.5/site-packages/asv/benchmark.py", line 773, in <module> commands[mode](args) File "/Library/Frameworks/Python.framework/Versions/3.5/lib/python3.5/site-packages/asv/benchmark.py", line 744, in main_run skip = benchmark.do_setup() File "/Library/Frameworks/Python.framework/Versions/3.5/lib/python3.5/site-packages/asv/benchmark.py", line 325, in do_setup setup(*self._current_params) TypeError: setup() takes 4 positional arguments but 5 were given ... Version information: Python 3.5 numpy 1.13.0 scipy 1.0.0.dev0+32d8e9e (current master) asv 0.3.dev1123+32d09e76 (current master) The problem seems to be that an extra parameter is passed from somewhere: ``setup() takes 4 positional arguments but 5 were given``. Checking those params shows that what's passed is ``name, ret_value, solver, None``. Unclear where the ``None`` is produced exactly. The ``BenchGlobal`` constructor which does not call ``super(BenchGlobal, self).__init__`` seems fragile. Changing the signature of setup by appending a ``none`` parameter fixes things, but that doesn't look quite right. Last time this was modified is in gh-6400. @andyfaff what do you think?
defect
optimize benchglobal broken as reported on the mailing list optimize benchglobal benchmarks are broken currently python runtests py bench optimize benchglobal fails on both my modified version to add stochasticbb testing and on the original scipy repository errors described in the pastebin above python runtests py bench optimize benchleastsquares works flawlessly python runtests py bench optimize benchsmoothunbounded works flawlessly traceback result type solver test function success de success basinh de basinh amgm failed failed failed failed ····· for parameters amgm success de traceback most recent call last file library frameworks python framework versions lib site packages asv benchmark py line in commands args file library frameworks python framework versions lib site packages asv benchmark py line in main run skip benchmark do setup file library frameworks python framework versions lib site packages asv benchmark py line in do setup setup self current params typeerror setup takes positional arguments but were given version information python numpy scipy current master asv current master the problem seems to be that an extra parameter is passed from somewhere setup takes positional arguments but were given checking those params shows that what s passed is name ret value solver none unclear where the none is produced exactly the benchglobal constructor which does not call super benchglobal self init seems fragile changing the signature of setup by appending a none parameter fixes things but that doesn t look quite right last time this was modified is in gh andyfaff what do you think
1
175,254
27,815,663,231
IssuesEvent
2023-03-18 16:56:59
Jade-ux/WomenTechConnect
https://api.github.com/repos/Jade-ux/WomenTechConnect
closed
Sign-up form - Design
design
Design and code separated into two cards but if you find it easier to go ahead and create it in the code straight away, that's great, feel free to assign both tickets to yourself Wireframes for inspiration: ![Image](https://user-images.githubusercontent.com/62659009/225906613-50df2a75-1447-495f-9824-a78b0b6f43f2.png)
1.0
Sign-up form - Design - Design and code separated into two cards but if you find it easier to go ahead and create it in the code straight away, that's great, feel free to assign both tickets to yourself Wireframes for inspiration: ![Image](https://user-images.githubusercontent.com/62659009/225906613-50df2a75-1447-495f-9824-a78b0b6f43f2.png)
non_defect
sign up form design design and code separated into two cards but if you find it easier to go ahead and create it in the code straight away that s great feel free to assign both tickets to yourself wireframes for inspiration
0
250,590
18,895,221,407
IssuesEvent
2021-11-15 17:07:36
apache/incubator-ponymail-foal
https://api.github.com/repos/apache/incubator-ponymail-foal
closed
What is this code trying to do?
bug documentation
https://github.com/apache/incubator-ponymail-foal/blob/6bbb4f49c99a22962afe7953b5fa06dfea341ec2/server/plugins/messages.py#L361-L364 AFAICT this will always result in the body being limited to BODY_MAXLEN + 1 (i.e. 201), as body_short is limited to SHORT_BODY_MAX_LEN+1 (i.e. 201) by the archiver. So what is the point of the shorten option? Why not just replace body with body_short?
1.0
What is this code trying to do? - https://github.com/apache/incubator-ponymail-foal/blob/6bbb4f49c99a22962afe7953b5fa06dfea341ec2/server/plugins/messages.py#L361-L364 AFAICT this will always result in the body being limited to BODY_MAXLEN + 1 (i.e. 201), as body_short is limited to SHORT_BODY_MAX_LEN+1 (i.e. 201) by the archiver. So what is the point of the shorten option? Why not just replace body with body_short?
non_defect
what is this code trying to do afaict this will always result in the body being limited to body maxlen i e as body short is limited to short body max len i e by the archiver so what is the point of the shorten option why not just replace body with body short
0
65,819
19,707,685,609
IssuesEvent
2022-01-13 00:30:55
jccastillo0007/eFacturaT
https://api.github.com/repos/jccastillo0007/eFacturaT
closed
CCP - misceláneos, no envía el número de registro ni residencia fiscal cuándo éstos se capturan,
defect
Estos campos se utilizan en traslados internacionales. Aquí lo que se necesita es enviarlos al XML, cuando se capturen. Por ejemplo, el conector como simplemente envía lo que incluye el archivo de texto, si lo hace: <cartaporte20:Ubicacion DistanciaRecorrida="1000.0" FechaHoraSalidaLlegada="2021-05-10T10:00:00" NombreRemitenteDestinatario="PREMIER AUTOCOUNTRY SA DE CV" NumRegIdTrib="123456" RFCRemitenteDestinatario="XEXX010101000" ResidenciaFiscal="AFG" TipoUbicacion="Destino"> En el caso de la plataforma, no lo hace. Simplemente no lo envía, y el SAT pues marca error cuando el rfc del origen o destino es XEXX010101000, lo cual indica que debe acompañarse de estos 2 datos que no está incluyendo.
1.0
CCP - misceláneos, no envía el número de registro ni residencia fiscal cuándo éstos se capturan, - Estos campos se utilizan en traslados internacionales. Aquí lo que se necesita es enviarlos al XML, cuando se capturen. Por ejemplo, el conector como simplemente envía lo que incluye el archivo de texto, si lo hace: <cartaporte20:Ubicacion DistanciaRecorrida="1000.0" FechaHoraSalidaLlegada="2021-05-10T10:00:00" NombreRemitenteDestinatario="PREMIER AUTOCOUNTRY SA DE CV" NumRegIdTrib="123456" RFCRemitenteDestinatario="XEXX010101000" ResidenciaFiscal="AFG" TipoUbicacion="Destino"> En el caso de la plataforma, no lo hace. Simplemente no lo envía, y el SAT pues marca error cuando el rfc del origen o destino es XEXX010101000, lo cual indica que debe acompañarse de estos 2 datos que no está incluyendo.
defect
ccp misceláneos no envía el número de registro ni residencia fiscal cuándo éstos se capturan estos campos se utilizan en traslados internacionales aquí lo que se necesita es enviarlos al xml cuando se capturen por ejemplo el conector como simplemente envía lo que incluye el archivo de texto si lo hace en el caso de la plataforma no lo hace simplemente no lo envía y el sat pues marca error cuando el rfc del origen o destino es lo cual indica que debe acompañarse de estos datos que no está incluyendo
1
214,791
7,276,787,907
IssuesEvent
2018-02-21 17:21:31
TylerConlee/slab
https://api.github.com/repos/TylerConlee/slab
closed
Add organization details to SLA ticket notifications
enhancement priority:normal
The organization that the ticket belongs to would be helpful for additional context on the ticket seen in the SLA notification.
1.0
Add organization details to SLA ticket notifications - The organization that the ticket belongs to would be helpful for additional context on the ticket seen in the SLA notification.
non_defect
add organization details to sla ticket notifications the organization that the ticket belongs to would be helpful for additional context on the ticket seen in the sla notification
0
519,994
15,076,874,988
IssuesEvent
2021-02-05 05:47:10
hassio-addons/addon-node-red
https://api.github.com/repos/hassio-addons/addon-node-red
closed
node-red-contrib-actionflows runtime errors (Deprecated Events)
priority-medium
# Problem/Motivation The node-red component "node-red-contrib-actionflows" was written prior to the availability of "flows:started" and uses an older method of flow setup that is no longer supported in NodeRED. This triggers runtime deprecation notices. ## Expected behavior No warnings, no indications on flow initiation. ## Actual behavior Debug panel display on deployment AND deprecation notice on Node RED startup with the node in the configuration: `[warn] [RED.events] Deprecated use of "nodes-started" event` ## Steps to reproduce 1. Add a flow with the node from "node-red-contrib-actionflows" in use (an inject -> node -> debug flow is fine for this purpose. 2. Display the Debug window in the side pane. 3. Deploy. ## Proposed changes Adopt version 2.0.4 with fix. PR to fix this deprecated event use has been merged by the maintainer. PR here: https://github.com/Steveorevo/node-red-contrib-actionflows/pull/15 Main git here: https://github.com/Steveorevo/node-red-contrib-actionflows
1.0
node-red-contrib-actionflows runtime errors (Deprecated Events) - # Problem/Motivation The node-red component "node-red-contrib-actionflows" was written prior to the availability of "flows:started" and uses an older method of flow setup that is no longer supported in NodeRED. This triggers runtime deprecation notices. ## Expected behavior No warnings, no indications on flow initiation. ## Actual behavior Debug panel display on deployment AND deprecation notice on Node RED startup with the node in the configuration: `[warn] [RED.events] Deprecated use of "nodes-started" event` ## Steps to reproduce 1. Add a flow with the node from "node-red-contrib-actionflows" in use (an inject -> node -> debug flow is fine for this purpose. 2. Display the Debug window in the side pane. 3. Deploy. ## Proposed changes Adopt version 2.0.4 with fix. PR to fix this deprecated event use has been merged by the maintainer. PR here: https://github.com/Steveorevo/node-red-contrib-actionflows/pull/15 Main git here: https://github.com/Steveorevo/node-red-contrib-actionflows
non_defect
node red contrib actionflows runtime errors deprecated events problem motivation the node red component node red contrib actionflows was written prior to the availability of flows started and uses an older method of flow setup that is no longer supported in nodered this triggers runtime deprecation notices expected behavior no warnings no indications on flow initiation actual behavior debug panel display on deployment and deprecation notice on node red startup with the node in the configuration deprecated use of nodes started event steps to reproduce add a flow with the node from node red contrib actionflows in use an inject node debug flow is fine for this purpose display the debug window in the side pane deploy proposed changes adopt version with fix pr to fix this deprecated event use has been merged by the maintainer pr here main git here
0
48,935
7,466,453,561
IssuesEvent
2018-04-02 10:43:00
scalameta/metals
https://api.github.com/repos/scalameta/metals
closed
Simplify installation
documentation installation
Currently, the installation steps are a sequence of several fairly fragile steps. https://github.com/scalameta/language-server/blob/master/BETA.md Any mistake could make nothing work, or half-work with no instructions on what's missing. We should strive for as simple installation as possible but avoid 100% automagical setup that makes it difficult to track down what went wrong. For sbt, I propose * addSbtPlugin * document how to enable semanticdb-scalac compiler plugin * `> metalsSetup` via sbt, downloads source jars, emit classpath and other stuff TODOs: * [ ] create sbt-metals plugin from project/MetalsPlugin.scala. We can depend on the plugin in the metals build by using `unmanagedSources in Compile ++= (baseDirectory.value.getParentFile / "sbt-metals").descedantsExcept("*.scala", ..).get`
1.0
Simplify installation - Currently, the installation steps are a sequence of several fairly fragile steps. https://github.com/scalameta/language-server/blob/master/BETA.md Any mistake could make nothing work, or half-work with no instructions on what's missing. We should strive for as simple installation as possible but avoid 100% automagical setup that makes it difficult to track down what went wrong. For sbt, I propose * addSbtPlugin * document how to enable semanticdb-scalac compiler plugin * `> metalsSetup` via sbt, downloads source jars, emit classpath and other stuff TODOs: * [ ] create sbt-metals plugin from project/MetalsPlugin.scala. We can depend on the plugin in the metals build by using `unmanagedSources in Compile ++= (baseDirectory.value.getParentFile / "sbt-metals").descedantsExcept("*.scala", ..).get`
non_defect
simplify installation currently the installation steps are a sequence of several fairly fragile steps any mistake could make nothing work or half work with no instructions on what s missing we should strive for as simple installation as possible but avoid automagical setup that makes it difficult to track down what went wrong for sbt i propose addsbtplugin document how to enable semanticdb scalac compiler plugin metalssetup via sbt downloads source jars emit classpath and other stuff todos create sbt metals plugin from project metalsplugin scala we can depend on the plugin in the metals build by using unmanagedsources in compile basedirectory value getparentfile sbt metals descedantsexcept scala get
0
319,509
23,775,980,719
IssuesEvent
2022-09-01 20:59:51
spacetelescope/drizzlepac
https://api.github.com/repos/spacetelescope/drizzlepac
closed
Need updates to DGEO warning message
Documentation
Hello! I'm working on a WFPC2 DrizzlePac ticket (INC0154128, for reference) and the user asked about the lengthy warning message below. I feel this needs to be updated for two reasons: 1) it provides users with an IRAF hedit command, and IRAF is deprecated, and 2) it should explain that for WFPC2 and other archival instruments, it is **fine** and **expected** to type 'c' and continue running drizzle. I don't believe that's clear at that moment. Thanks!! > > A 'DGEOFILE' keyword is present in the primary header but 'NPOLFILE' keyword was not found. > This version of the software uses a new format for the residual distortion DGEO files. > Please consult the instrument web pages for which reference files to download. > A small (new style) dgeofile is needed ('_npl.fits' extension) and possibly a > detector to image correction file ('_d2i.fits' extension). > The names of these files must be added to the primary header either using the task XXXX > or manually, for example: > > hedit u2pq0505t_c0m.fits[0] npolfile fname_npl.fits add+ > hedit u2pq0505t_c0m.fits[0] d2imfile fname_d2i.fits add+ > > where fname_npl.fits is the name of the new style dgeo file and fname_d2i.fits is > the name of the detector to image correction. After adding these keywords to the > primary header, updatewcs must be run to update the science files: > > from stwcs import updatewcs > updatewcs.updatewcs("u2pq0505t_c0m.fits") > > Alternatively you may choose to run astrodrizzle without DGEO and detector to image correction. > > To stop astrodrizzle and update the dgeo files, type 'q'. > To continue running astrodrizzle without the non-polynomial distortion correction, type 'c':
1.0
Need updates to DGEO warning message - Hello! I'm working on a WFPC2 DrizzlePac ticket (INC0154128, for reference) and the user asked about the lengthy warning message below. I feel this needs to be updated for two reasons: 1) it provides users with an IRAF hedit command, and IRAF is deprecated, and 2) it should explain that for WFPC2 and other archival instruments, it is **fine** and **expected** to type 'c' and continue running drizzle. I don't believe that's clear at that moment. Thanks!! > > A 'DGEOFILE' keyword is present in the primary header but 'NPOLFILE' keyword was not found. > This version of the software uses a new format for the residual distortion DGEO files. > Please consult the instrument web pages for which reference files to download. > A small (new style) dgeofile is needed ('_npl.fits' extension) and possibly a > detector to image correction file ('_d2i.fits' extension). > The names of these files must be added to the primary header either using the task XXXX > or manually, for example: > > hedit u2pq0505t_c0m.fits[0] npolfile fname_npl.fits add+ > hedit u2pq0505t_c0m.fits[0] d2imfile fname_d2i.fits add+ > > where fname_npl.fits is the name of the new style dgeo file and fname_d2i.fits is > the name of the detector to image correction. After adding these keywords to the > primary header, updatewcs must be run to update the science files: > > from stwcs import updatewcs > updatewcs.updatewcs("u2pq0505t_c0m.fits") > > Alternatively you may choose to run astrodrizzle without DGEO and detector to image correction. > > To stop astrodrizzle and update the dgeo files, type 'q'. > To continue running astrodrizzle without the non-polynomial distortion correction, type 'c':
non_defect
need updates to dgeo warning message hello i m working on a drizzlepac ticket for reference and the user asked about the lengthy warning message below i feel this needs to be updated for two reasons it provides users with an iraf hedit command and iraf is deprecated and it should explain that for and other archival instruments it is fine and expected to type c and continue running drizzle i don t believe that s clear at that moment thanks a dgeofile keyword is present in the primary header but npolfile keyword was not found this version of the software uses a new format for the residual distortion dgeo files please consult the instrument web pages for which reference files to download a small new style dgeofile is needed npl fits extension and possibly a detector to image correction file fits extension the names of these files must be added to the primary header either using the task xxxx or manually for example hedit fits npolfile fname npl fits add hedit fits fname fits add where fname npl fits is the name of the new style dgeo file and fname fits is the name of the detector to image correction after adding these keywords to the primary header updatewcs must be run to update the science files from stwcs import updatewcs updatewcs updatewcs fits alternatively you may choose to run astrodrizzle without dgeo and detector to image correction to stop astrodrizzle and update the dgeo files type q to continue running astrodrizzle without the non polynomial distortion correction type c
0
103,135
11,340,124,514
IssuesEvent
2020-01-23 05:09:17
wayexists02/hanyang-erica-robot-programming
https://api.github.com/repos/wayexists02/hanyang-erica-robot-programming
reopened
코드론 실행법
documentation
jylee 핫스팟(...)에 서버와 코드론을 연결한 후, 서버와 코드론에서 코드를 실행한다. ## 서버의 실행 코드론을 실행하기 앞서서 먼저 실행한다. ```roslaunch codrone_alpha launch.launch``` ## 코드론의 실행 ```roslaunch codrone_alpha_pi launch.launch```
1.0
코드론 실행법 - jylee 핫스팟(...)에 서버와 코드론을 연결한 후, 서버와 코드론에서 코드를 실행한다. ## 서버의 실행 코드론을 실행하기 앞서서 먼저 실행한다. ```roslaunch codrone_alpha launch.launch``` ## 코드론의 실행 ```roslaunch codrone_alpha_pi launch.launch```
non_defect
코드론 실행법 jylee 핫스팟 에 서버와 코드론을 연결한 후 서버와 코드론에서 코드를 실행한다 서버의 실행 코드론을 실행하기 앞서서 먼저 실행한다 roslaunch codrone alpha launch launch 코드론의 실행 roslaunch codrone alpha pi launch launch
0
4,403
22,617,321,211
IssuesEvent
2022-06-30 00:20:29
aws/aws-sam-cli
https://api.github.com/repos/aws/aws-sam-cli
closed
`sam sync` does not support custom bucket names
type/ux type/feature area/sam-config area/sync maintainer/need-followup area/accelerate
### Description: I don't use the default SAM bucket, I have my own. `sam sync` does not seem to support this. ### Steps to reproduce: Do `sam init` and create the zip Python 3.9 "Hello World" template. Create the following samconfig.toml ```toml version = 0.1 [default] [default.deploy] [default.deploy.parameters] stack_name = "sam-test" s3_bucket = "mybucket" s3_prefix = "sam-test" region = "us-west-2" capabilities = "CAPABILITY_IAM" ``` Run `sam build && sam deploy`, which succeeds. ### Observed result: `sam sync --stack-name sam-test` gives the following output. You can see it's attempting to use the default managed SAM bucket. ``` 2021-12-17 11:40:14,807 | Telemetry endpoint configured to be https://aws-serverless-tools-telemetry.us-west-2.amazonaws.com/metrics 2021-12-17 11:40:14,812 | Telemetry endpoint configured to be https://aws-serverless-tools-telemetry.us-west-2.amazonaws.com/metrics 2021-12-17 11:40:14,812 | Sending Telemetry: {'metrics': [{'templateWarning': {'requestId': '5e92f8cb-75e3-4793-81f8-faee808f01a7', 'installationId': '1ef32602-7319-4d1a-bc65-fb2419c3fe35', 'sessionId': 'eeb5b278-0298-446b-9bcc-43424c2cd44d', 'executionEnvironment': 'CLI', 'ci': False, 'pyversion': '3.8.12', 'samcliVersion': '1.36.0', 'awsProfileProvided': False, 'debugFlagProvided': True, 'region': '', 'warningName': 'CodeDeployWarning', 'warningCount': 0}}]} 2021-12-17 11:40:15,017 | Telemetry response: 200 2021-12-17 11:40:15,018 | Sending Telemetry: {'metrics': [{'templateWarning': {'requestId': 'd0f3bfd9-c6d7-40db-9c8b-337bf8efcd98', 'installationId': '1ef32602-7319-4d1a-bc65-fb2419c3fe35', 'sessionId': 'eeb5b278-0298-446b-9bcc-43424c2cd44d', 'executionEnvironment': 'CLI', 'ci': False, 'pyversion': '3.8.12', 'samcliVersion': '1.36.0', 'awsProfileProvided': False, 'debugFlagProvided': True, 'region': '', 'warningName': 'CodeDeployConditionWarning', 'warningCount': 0}}]} 2021-12-17 11:40:15,283 | Telemetry response: 200 2021-12-17 11:40:15,284 | Using config file: samconfig.toml, config environment: default 2021-12-17 11:40:15,284 | Expand command line arguments to: 2021-12-17 11:40:15,284 | --template_file=/Users/luhn/Code/audit/test/template.yaml --stack_name=sam-test --dependency_layer --capabilities=('CAPABILITY_NAMED_IAM', 'CAPABILITY_AUTO_EXPAND') Managed S3 bucket: aws-sam-cli-managed-default-samclisourcebucket-1aupim17uw7m6 Default capabilities applied: ('CAPABILITY_NAMED_IAM', 'CAPABILITY_AUTO_EXPAND') To override with customized capabilities, use --capabilities flag or set it in samconfig.toml 2021-12-17 11:40:16,112 | Using build directory as .aws-sam/auto-dependency-layer 2021-12-17 11:40:16,112 | Using build directory as .aws-sam/auto-dependency-layer This feature is currently in beta. Visit the docs page to learn more about the AWS Beta terms https://aws.amazon.com/service-terms/. The SAM CLI will use the AWS Lambda, Amazon API Gateway, and AWS StepFunctions APIs to upload your code without performing a CloudFormation deployment. This will cause drift in your CloudFormation stack. **The sync command should only be used against a development stack**. Confirm that you are synchronizing a development stack and want to turn on beta features. Enter Y to proceed with the command, or enter N to cancel: [y/N]: 2021-12-17 11:40:17,467 |  Experimental features are enabled for this session. Visit the docs page to learn more about the AWS Beta terms https://aws.amazon.com/service-terms/.  2021-12-17 11:40:17,477 | No Parameters detected in the template 2021-12-17 11:40:17,499 | 2 stacks found in the template 2021-12-17 11:40:17,499 | No Parameters detected in the template 2021-12-17 11:40:17,510 | 2 resources found in the stack 2021-12-17 11:40:17,510 | No Parameters detected in the template 2021-12-17 11:40:17,519 | Found Serverless function with name='HelloWorldFunction' and CodeUri='hello_world/' 2021-12-17 11:40:17,519 | --base-dir is not presented, adjusting uri hello_world/ relative to /Users/luhn/Code/audit/test/template.yaml 2021-12-17 11:40:17,519 | No Parameters detected in the template 2021-12-17 11:40:17,538 | Executing the build using build context. 2021-12-17 11:40:17,538 | Instantiating build definitions 2021-12-17 11:40:17,540 | Same function build definition found, adding function (Previous: BuildDefinition(python3.9, /Users/luhn/Code/audit/test/hello_world, Zip, , d23e058e-cbff-4bce-85b2-09954cf33d29, {}, {}, x86_64, []), Current: BuildDefinition(python3.9, /Users/luhn/Code/audit/test/hello_world, Zip, , 85a07967-200c-4a31-81df-7700103e6ad7, {}, {}, x86_64, []), Function: Function(name='HelloWorldFunction', functionname='HelloWorldFunction', runtime='python3.9', memory=None, timeout=3, handler='app.lambda_handler', imageuri=None, packagetype='Zip', imageconfig=None, codeuri='/Users/luhn/Code/audit/test/hello_world', environment=None, rolearn=None, layers=[], events={'HelloWorld': {'Type': 'Api', 'Properties': {'Path': '/hello', 'Method': 'get', 'RestApiId': 'ServerlessRestApi'}}}, metadata=None, inlinecode=None, codesign_config_arn=None, architectures=['x86_64'], stack_path='')) 2021-12-17 11:40:17,541 | Async execution started 2021-12-17 11:40:17,541 | Invoking function functools.partial(<bound method CachedOrIncrementalBuildStrategyWrapper.build_single_function_definition of <samcli.lib.build.build_strategy.CachedOrIncrementalBuildStrategyWrapper object at 0x1056eb3d0>>, <samcli.lib.build.build_graph.FunctionBuildDefinition object at 0x1053468e0>) 2021-12-17 11:40:17,541 | Running incremental build for runtime python3.9 for build definition d23e058e-cbff-4bce-85b2-09954cf33d29 2021-12-17 11:40:17,541 | Waiting for async results 2021-12-17 11:40:17,541 | Manifest is not changed for d23e058e-cbff-4bce-85b2-09954cf33d29, running incremental build 2021-12-17 11:40:17,541 | Building codeuri: /Users/luhn/Code/audit/test/hello_world runtime: python3.9 metadata: {} architecture: x86_64 functions: ['HelloWorldFunction'] 2021-12-17 11:40:17,541 | Building to following folder /Users/luhn/Code/audit/test/.aws-sam/auto-dependency-layer/HelloWorldFunction 2021-12-17 11:40:17,542 | Loading workflow module 'aws_lambda_builders.workflows' 2021-12-17 11:40:17,546 | Registering workflow 'PythonPipBuilder' with capability 'Capability(language='python', dependency_manager='pip', application_framework=None)' 2021-12-17 11:40:17,548 | Registering workflow 'NodejsNpmBuilder' with capability 'Capability(language='nodejs', dependency_manager='npm', application_framework=None)' 2021-12-17 11:40:17,549 | Registering workflow 'RubyBundlerBuilder' with capability 'Capability(language='ruby', dependency_manager='bundler', application_framework=None)' 2021-12-17 11:40:17,551 | Registering workflow 'GoDepBuilder' with capability 'Capability(language='go', dependency_manager='dep', application_framework=None)' 2021-12-17 11:40:17,553 | Registering workflow 'GoModulesBuilder' with capability 'Capability(language='go', dependency_manager='modules', application_framework=None)' 2021-12-17 11:40:17,555 | Registering workflow 'JavaGradleWorkflow' with capability 'Capability(language='java', dependency_manager='gradle', application_framework=None)' 2021-12-17 11:40:17,556 | Registering workflow 'JavaMavenWorkflow' with capability 'Capability(language='java', dependency_manager='maven', application_framework=None)' 2021-12-17 11:40:17,558 | Registering workflow 'DotnetCliPackageBuilder' with capability 'Capability(language='dotnet', dependency_manager='cli-package', application_framework=None)' 2021-12-17 11:40:17,559 | Registering workflow 'CustomMakeBuilder' with capability 'Capability(language='provided', dependency_manager=None, application_framework=None)' 2021-12-17 11:40:17,559 | Found workflow 'PythonPipBuilder' to support capabilities 'Capability(language='python', dependency_manager='pip', application_framework=None)' 2021-12-17 11:40:17,626 | Running workflow 'PythonPipBuilder' 2021-12-17 11:40:17,627 | Running PythonPipBuilder:CopySource 2021-12-17 11:40:17,629 | PythonPipBuilder:CopySource succeeded 2021-12-17 11:40:17,629 | Async execution completed 2021-12-17 11:40:17,630 | Auto creating dependency layer for each function resource into a nested stack 2021-12-17 11:40:17,630 | No Parameters detected in the template 2021-12-17 11:40:17,636 | 2 resources found in the stack sam-test 2021-12-17 11:40:17,636 | No Parameters detected in the template 2021-12-17 11:40:17,641 | Found Serverless function with name='HelloWorldFunction' and CodeUri='.aws-sam/auto-dependency-layer/HelloWorldFunction' 2021-12-17 11:40:17,641 | --base-dir is not presented, adjusting uri .aws-sam/auto-dependency-layer/HelloWorldFunction relative to /Users/luhn/Code/audit/test/template.yaml Build Succeeded Built Artifacts : .aws-sam/auto-dependency-layer Built Template : .aws-sam/auto-dependency-layer/template.yaml Commands you can use next ========================= [*] Invoke Function: sam local invoke -t .aws-sam/auto-dependency-layer/template.yaml [*] Test Function in the Cloud: sam sync --stack-name {stack-name} --watch [*] Deploy: sam deploy --guided --template-file .aws-sam/auto-dependency-layer/template.yaml 2021-12-17 11:40:17,667 | Executing the packaging using package context. 2021-12-17 11:40:18,030 | Unable to export Traceback (most recent call last): File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/samcli/lib/package/s3_uploader.py", line 114, in upload future.result() File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/s3transfer/futures.py", line 106, in result return self._coordinator.result() File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/s3transfer/futures.py", line 265, in result raise self._exception File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/s3transfer/tasks.py", line 126, in __call__ return self._execute_main(kwargs) File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/s3transfer/tasks.py", line 150, in _execute_main return_value = self._main(**kwargs) File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/s3transfer/upload.py", line 694, in _main client.put_object(Bucket=bucket, Key=key, Body=body, **extra_args) File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/botocore/client.py", line 391, in _api_call return self._make_api_call(operation_name, kwargs) File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/botocore/client.py", line 719, in _make_api_call raise error_class(parsed_response, operation_name) botocore.errorfactory.NoSuchBucket: An error occurred (NoSuchBucket) when calling the PutObject operation: The specified bucket does not exist The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/samcli/lib/package/packageable_resources.py", line 126, in export self.do_export(resource_id, resource_dict, parent_dir) File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/samcli/lib/package/packageable_resources.py", line 148, in do_export uploaded_url = upload_local_artifacts( File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/samcli/lib/package/utils.py", line 171, in upload_local_artifacts return zip_and_upload(local_path, uploader, extension) File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/samcli/lib/package/utils.py", line 189, in zip_and_upload return uploader.upload_with_dedup(zip_file, precomputed_md5=md5_hash, extension=extension) File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/samcli/lib/package/s3_uploader.py", line 143, in upload_with_dedup return self.upload(file_name, remote_path) File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/samcli/lib/package/s3_uploader.py", line 121, in upload raise NoSuchBucketError(bucket_name=self.bucket_name) from ex samcli.commands.package.exceptions.NoSuchBucketError: S3 Bucket does not exist. 2021-12-17 11:40:18,033 | Sending Telemetry: {'metrics': [{'commandRunExperimental': {'requestId': '2898b15c-f378-4219-b192-da75e8d8e59d', 'installationId': '1ef32602-7319-4d1a-bc65-fb2419c3fe35', 'sessionId': 'eeb5b278-0298-446b-9bcc-43424c2cd44d', 'executionEnvironment': 'CLI', 'ci': False, 'pyversion': '3.8.12', 'samcliVersion': '1.36.0', 'awsProfileProvided': False, 'debugFlagProvided': True, 'region': '', 'commandName': 'sam sync', 'metricSpecificAttributes': {'experimentalAccelerate': True, 'experimentalAll': False}, 'duration': 3225, 'exitReason': 'ExportFailedError', 'exitCode': 1}}]} 2021-12-17 11:40:18,278 | Telemetry response: 200 Error: Unable to upload artifact HelloWorldFunction referenced by CodeUri parameter of HelloWorldFunction resource. S3 Bucket does not exist. ``` ### Expected result: I would expect a) sync to honor the settings in samconfig.toml or b) a CLI flag to set the S3 bucket name. ### Additional environment details (Ex: Windows, Mac, Amazon Linux etc) 1. OS: Mac OS Monterey 2. If using SAM CLI, `sam --version`: `SAM CLI, version 1.36.0` 3. AWS region: us-west-2
True
`sam sync` does not support custom bucket names - ### Description: I don't use the default SAM bucket, I have my own. `sam sync` does not seem to support this. ### Steps to reproduce: Do `sam init` and create the zip Python 3.9 "Hello World" template. Create the following samconfig.toml ```toml version = 0.1 [default] [default.deploy] [default.deploy.parameters] stack_name = "sam-test" s3_bucket = "mybucket" s3_prefix = "sam-test" region = "us-west-2" capabilities = "CAPABILITY_IAM" ``` Run `sam build && sam deploy`, which succeeds. ### Observed result: `sam sync --stack-name sam-test` gives the following output. You can see it's attempting to use the default managed SAM bucket. ``` 2021-12-17 11:40:14,807 | Telemetry endpoint configured to be https://aws-serverless-tools-telemetry.us-west-2.amazonaws.com/metrics 2021-12-17 11:40:14,812 | Telemetry endpoint configured to be https://aws-serverless-tools-telemetry.us-west-2.amazonaws.com/metrics 2021-12-17 11:40:14,812 | Sending Telemetry: {'metrics': [{'templateWarning': {'requestId': '5e92f8cb-75e3-4793-81f8-faee808f01a7', 'installationId': '1ef32602-7319-4d1a-bc65-fb2419c3fe35', 'sessionId': 'eeb5b278-0298-446b-9bcc-43424c2cd44d', 'executionEnvironment': 'CLI', 'ci': False, 'pyversion': '3.8.12', 'samcliVersion': '1.36.0', 'awsProfileProvided': False, 'debugFlagProvided': True, 'region': '', 'warningName': 'CodeDeployWarning', 'warningCount': 0}}]} 2021-12-17 11:40:15,017 | Telemetry response: 200 2021-12-17 11:40:15,018 | Sending Telemetry: {'metrics': [{'templateWarning': {'requestId': 'd0f3bfd9-c6d7-40db-9c8b-337bf8efcd98', 'installationId': '1ef32602-7319-4d1a-bc65-fb2419c3fe35', 'sessionId': 'eeb5b278-0298-446b-9bcc-43424c2cd44d', 'executionEnvironment': 'CLI', 'ci': False, 'pyversion': '3.8.12', 'samcliVersion': '1.36.0', 'awsProfileProvided': False, 'debugFlagProvided': True, 'region': '', 'warningName': 'CodeDeployConditionWarning', 'warningCount': 0}}]} 2021-12-17 11:40:15,283 | Telemetry response: 200 2021-12-17 11:40:15,284 | Using config file: samconfig.toml, config environment: default 2021-12-17 11:40:15,284 | Expand command line arguments to: 2021-12-17 11:40:15,284 | --template_file=/Users/luhn/Code/audit/test/template.yaml --stack_name=sam-test --dependency_layer --capabilities=('CAPABILITY_NAMED_IAM', 'CAPABILITY_AUTO_EXPAND') Managed S3 bucket: aws-sam-cli-managed-default-samclisourcebucket-1aupim17uw7m6 Default capabilities applied: ('CAPABILITY_NAMED_IAM', 'CAPABILITY_AUTO_EXPAND') To override with customized capabilities, use --capabilities flag or set it in samconfig.toml 2021-12-17 11:40:16,112 | Using build directory as .aws-sam/auto-dependency-layer 2021-12-17 11:40:16,112 | Using build directory as .aws-sam/auto-dependency-layer This feature is currently in beta. Visit the docs page to learn more about the AWS Beta terms https://aws.amazon.com/service-terms/. The SAM CLI will use the AWS Lambda, Amazon API Gateway, and AWS StepFunctions APIs to upload your code without performing a CloudFormation deployment. This will cause drift in your CloudFormation stack. **The sync command should only be used against a development stack**. Confirm that you are synchronizing a development stack and want to turn on beta features. Enter Y to proceed with the command, or enter N to cancel: [y/N]: 2021-12-17 11:40:17,467 |  Experimental features are enabled for this session. Visit the docs page to learn more about the AWS Beta terms https://aws.amazon.com/service-terms/.  2021-12-17 11:40:17,477 | No Parameters detected in the template 2021-12-17 11:40:17,499 | 2 stacks found in the template 2021-12-17 11:40:17,499 | No Parameters detected in the template 2021-12-17 11:40:17,510 | 2 resources found in the stack 2021-12-17 11:40:17,510 | No Parameters detected in the template 2021-12-17 11:40:17,519 | Found Serverless function with name='HelloWorldFunction' and CodeUri='hello_world/' 2021-12-17 11:40:17,519 | --base-dir is not presented, adjusting uri hello_world/ relative to /Users/luhn/Code/audit/test/template.yaml 2021-12-17 11:40:17,519 | No Parameters detected in the template 2021-12-17 11:40:17,538 | Executing the build using build context. 2021-12-17 11:40:17,538 | Instantiating build definitions 2021-12-17 11:40:17,540 | Same function build definition found, adding function (Previous: BuildDefinition(python3.9, /Users/luhn/Code/audit/test/hello_world, Zip, , d23e058e-cbff-4bce-85b2-09954cf33d29, {}, {}, x86_64, []), Current: BuildDefinition(python3.9, /Users/luhn/Code/audit/test/hello_world, Zip, , 85a07967-200c-4a31-81df-7700103e6ad7, {}, {}, x86_64, []), Function: Function(name='HelloWorldFunction', functionname='HelloWorldFunction', runtime='python3.9', memory=None, timeout=3, handler='app.lambda_handler', imageuri=None, packagetype='Zip', imageconfig=None, codeuri='/Users/luhn/Code/audit/test/hello_world', environment=None, rolearn=None, layers=[], events={'HelloWorld': {'Type': 'Api', 'Properties': {'Path': '/hello', 'Method': 'get', 'RestApiId': 'ServerlessRestApi'}}}, metadata=None, inlinecode=None, codesign_config_arn=None, architectures=['x86_64'], stack_path='')) 2021-12-17 11:40:17,541 | Async execution started 2021-12-17 11:40:17,541 | Invoking function functools.partial(<bound method CachedOrIncrementalBuildStrategyWrapper.build_single_function_definition of <samcli.lib.build.build_strategy.CachedOrIncrementalBuildStrategyWrapper object at 0x1056eb3d0>>, <samcli.lib.build.build_graph.FunctionBuildDefinition object at 0x1053468e0>) 2021-12-17 11:40:17,541 | Running incremental build for runtime python3.9 for build definition d23e058e-cbff-4bce-85b2-09954cf33d29 2021-12-17 11:40:17,541 | Waiting for async results 2021-12-17 11:40:17,541 | Manifest is not changed for d23e058e-cbff-4bce-85b2-09954cf33d29, running incremental build 2021-12-17 11:40:17,541 | Building codeuri: /Users/luhn/Code/audit/test/hello_world runtime: python3.9 metadata: {} architecture: x86_64 functions: ['HelloWorldFunction'] 2021-12-17 11:40:17,541 | Building to following folder /Users/luhn/Code/audit/test/.aws-sam/auto-dependency-layer/HelloWorldFunction 2021-12-17 11:40:17,542 | Loading workflow module 'aws_lambda_builders.workflows' 2021-12-17 11:40:17,546 | Registering workflow 'PythonPipBuilder' with capability 'Capability(language='python', dependency_manager='pip', application_framework=None)' 2021-12-17 11:40:17,548 | Registering workflow 'NodejsNpmBuilder' with capability 'Capability(language='nodejs', dependency_manager='npm', application_framework=None)' 2021-12-17 11:40:17,549 | Registering workflow 'RubyBundlerBuilder' with capability 'Capability(language='ruby', dependency_manager='bundler', application_framework=None)' 2021-12-17 11:40:17,551 | Registering workflow 'GoDepBuilder' with capability 'Capability(language='go', dependency_manager='dep', application_framework=None)' 2021-12-17 11:40:17,553 | Registering workflow 'GoModulesBuilder' with capability 'Capability(language='go', dependency_manager='modules', application_framework=None)' 2021-12-17 11:40:17,555 | Registering workflow 'JavaGradleWorkflow' with capability 'Capability(language='java', dependency_manager='gradle', application_framework=None)' 2021-12-17 11:40:17,556 | Registering workflow 'JavaMavenWorkflow' with capability 'Capability(language='java', dependency_manager='maven', application_framework=None)' 2021-12-17 11:40:17,558 | Registering workflow 'DotnetCliPackageBuilder' with capability 'Capability(language='dotnet', dependency_manager='cli-package', application_framework=None)' 2021-12-17 11:40:17,559 | Registering workflow 'CustomMakeBuilder' with capability 'Capability(language='provided', dependency_manager=None, application_framework=None)' 2021-12-17 11:40:17,559 | Found workflow 'PythonPipBuilder' to support capabilities 'Capability(language='python', dependency_manager='pip', application_framework=None)' 2021-12-17 11:40:17,626 | Running workflow 'PythonPipBuilder' 2021-12-17 11:40:17,627 | Running PythonPipBuilder:CopySource 2021-12-17 11:40:17,629 | PythonPipBuilder:CopySource succeeded 2021-12-17 11:40:17,629 | Async execution completed 2021-12-17 11:40:17,630 | Auto creating dependency layer for each function resource into a nested stack 2021-12-17 11:40:17,630 | No Parameters detected in the template 2021-12-17 11:40:17,636 | 2 resources found in the stack sam-test 2021-12-17 11:40:17,636 | No Parameters detected in the template 2021-12-17 11:40:17,641 | Found Serverless function with name='HelloWorldFunction' and CodeUri='.aws-sam/auto-dependency-layer/HelloWorldFunction' 2021-12-17 11:40:17,641 | --base-dir is not presented, adjusting uri .aws-sam/auto-dependency-layer/HelloWorldFunction relative to /Users/luhn/Code/audit/test/template.yaml Build Succeeded Built Artifacts : .aws-sam/auto-dependency-layer Built Template : .aws-sam/auto-dependency-layer/template.yaml Commands you can use next ========================= [*] Invoke Function: sam local invoke -t .aws-sam/auto-dependency-layer/template.yaml [*] Test Function in the Cloud: sam sync --stack-name {stack-name} --watch [*] Deploy: sam deploy --guided --template-file .aws-sam/auto-dependency-layer/template.yaml 2021-12-17 11:40:17,667 | Executing the packaging using package context. 2021-12-17 11:40:18,030 | Unable to export Traceback (most recent call last): File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/samcli/lib/package/s3_uploader.py", line 114, in upload future.result() File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/s3transfer/futures.py", line 106, in result return self._coordinator.result() File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/s3transfer/futures.py", line 265, in result raise self._exception File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/s3transfer/tasks.py", line 126, in __call__ return self._execute_main(kwargs) File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/s3transfer/tasks.py", line 150, in _execute_main return_value = self._main(**kwargs) File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/s3transfer/upload.py", line 694, in _main client.put_object(Bucket=bucket, Key=key, Body=body, **extra_args) File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/botocore/client.py", line 391, in _api_call return self._make_api_call(operation_name, kwargs) File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/botocore/client.py", line 719, in _make_api_call raise error_class(parsed_response, operation_name) botocore.errorfactory.NoSuchBucket: An error occurred (NoSuchBucket) when calling the PutObject operation: The specified bucket does not exist The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/samcli/lib/package/packageable_resources.py", line 126, in export self.do_export(resource_id, resource_dict, parent_dir) File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/samcli/lib/package/packageable_resources.py", line 148, in do_export uploaded_url = upload_local_artifacts( File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/samcli/lib/package/utils.py", line 171, in upload_local_artifacts return zip_and_upload(local_path, uploader, extension) File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/samcli/lib/package/utils.py", line 189, in zip_and_upload return uploader.upload_with_dedup(zip_file, precomputed_md5=md5_hash, extension=extension) File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/samcli/lib/package/s3_uploader.py", line 143, in upload_with_dedup return self.upload(file_name, remote_path) File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/samcli/lib/package/s3_uploader.py", line 121, in upload raise NoSuchBucketError(bucket_name=self.bucket_name) from ex samcli.commands.package.exceptions.NoSuchBucketError: S3 Bucket does not exist. 2021-12-17 11:40:18,033 | Sending Telemetry: {'metrics': [{'commandRunExperimental': {'requestId': '2898b15c-f378-4219-b192-da75e8d8e59d', 'installationId': '1ef32602-7319-4d1a-bc65-fb2419c3fe35', 'sessionId': 'eeb5b278-0298-446b-9bcc-43424c2cd44d', 'executionEnvironment': 'CLI', 'ci': False, 'pyversion': '3.8.12', 'samcliVersion': '1.36.0', 'awsProfileProvided': False, 'debugFlagProvided': True, 'region': '', 'commandName': 'sam sync', 'metricSpecificAttributes': {'experimentalAccelerate': True, 'experimentalAll': False}, 'duration': 3225, 'exitReason': 'ExportFailedError', 'exitCode': 1}}]} 2021-12-17 11:40:18,278 | Telemetry response: 200 Error: Unable to upload artifact HelloWorldFunction referenced by CodeUri parameter of HelloWorldFunction resource. S3 Bucket does not exist. ``` ### Expected result: I would expect a) sync to honor the settings in samconfig.toml or b) a CLI flag to set the S3 bucket name. ### Additional environment details (Ex: Windows, Mac, Amazon Linux etc) 1. OS: Mac OS Monterey 2. If using SAM CLI, `sam --version`: `SAM CLI, version 1.36.0` 3. AWS region: us-west-2
non_defect
sam sync does not support custom bucket names description i don t use the default sam bucket i have my own sam sync does not seem to support this steps to reproduce do sam init and create the zip python hello world template create the following samconfig toml toml version stack name sam test bucket mybucket prefix sam test region us west capabilities capability iam run sam build sam deploy which succeeds observed result sam sync stack name sam test gives the following output you can see it s attempting to use the default managed sam bucket telemetry endpoint configured to be telemetry endpoint configured to be sending telemetry metrics telemetry response sending telemetry metrics telemetry response using config file samconfig toml config environment default expand command line arguments to template file users luhn code audit test template yaml stack name sam test dependency layer capabilities capability named iam capability auto expand managed bucket aws sam cli managed default samclisourcebucket default capabilities applied capability named iam capability auto expand to override with customized capabilities use capabilities flag or set it in samconfig toml using build directory as aws sam auto dependency layer using build directory as aws sam auto dependency layer this feature is currently in beta visit the docs page to learn more about the aws beta terms the sam cli will use the aws lambda amazon api gateway and aws stepfunctions apis to upload your code without performing a cloudformation deployment this will cause drift in your cloudformation stack the sync command should only be used against a development stack confirm that you are synchronizing a development stack and want to turn on beta features enter y to proceed with the command or enter n to cancel  experimental features are enabled for this session visit the docs page to learn more about the aws beta terms  no parameters detected in the template stacks found in the template no parameters detected in the template resources found in the stack no parameters detected in the template found serverless function with name helloworldfunction and codeuri hello world base dir is not presented adjusting uri hello world relative to users luhn code audit test template yaml no parameters detected in the template executing the build using build context instantiating build definitions same function build definition found adding function previous builddefinition users luhn code audit test hello world zip cbff current builddefinition users luhn code audit test hello world zip function function name helloworldfunction functionname helloworldfunction runtime memory none timeout handler app lambda handler imageuri none packagetype zip imageconfig none codeuri users luhn code audit test hello world environment none rolearn none layers events helloworld type api properties path hello method get restapiid serverlessrestapi metadata none inlinecode none codesign config arn none architectures stack path async execution started invoking function functools partial running incremental build for runtime for build definition cbff waiting for async results manifest is not changed for cbff running incremental build building codeuri users luhn code audit test hello world runtime metadata architecture functions building to following folder users luhn code audit test aws sam auto dependency layer helloworldfunction loading workflow module aws lambda builders workflows registering workflow pythonpipbuilder with capability capability language python dependency manager pip application framework none registering workflow nodejsnpmbuilder with capability capability language nodejs dependency manager npm application framework none registering workflow rubybundlerbuilder with capability capability language ruby dependency manager bundler application framework none registering workflow godepbuilder with capability capability language go dependency manager dep application framework none registering workflow gomodulesbuilder with capability capability language go dependency manager modules application framework none registering workflow javagradleworkflow with capability capability language java dependency manager gradle application framework none registering workflow javamavenworkflow with capability capability language java dependency manager maven application framework none registering workflow dotnetclipackagebuilder with capability capability language dotnet dependency manager cli package application framework none registering workflow custommakebuilder with capability capability language provided dependency manager none application framework none found workflow pythonpipbuilder to support capabilities capability language python dependency manager pip application framework none running workflow pythonpipbuilder running pythonpipbuilder copysource pythonpipbuilder copysource succeeded async execution completed auto creating dependency layer for each function resource into a nested stack no parameters detected in the template resources found in the stack sam test no parameters detected in the template found serverless function with name helloworldfunction and codeuri aws sam auto dependency layer helloworldfunction base dir is not presented adjusting uri aws sam auto dependency layer helloworldfunction relative to users luhn code audit test template yaml build succeeded built artifacts aws sam auto dependency layer built template aws sam auto dependency layer template yaml commands you can use next invoke function sam local invoke t aws sam auto dependency layer template yaml test function in the cloud sam sync stack name stack name watch deploy sam deploy guided template file aws sam auto dependency layer template yaml executing the packaging using package context unable to export traceback most recent call last file opt homebrew cellar aws sam cli libexec lib site packages samcli lib package uploader py line in upload future result file opt homebrew cellar aws sam cli libexec lib site packages futures py line in result return self coordinator result file opt homebrew cellar aws sam cli libexec lib site packages futures py line in result raise self exception file opt homebrew cellar aws sam cli libexec lib site packages tasks py line in call return self execute main kwargs file opt homebrew cellar aws sam cli libexec lib site packages tasks py line in execute main return value self main kwargs file opt homebrew cellar aws sam cli libexec lib site packages upload py line in main client put object bucket bucket key key body body extra args file opt homebrew cellar aws sam cli libexec lib site packages botocore client py line in api call return self make api call operation name kwargs file opt homebrew cellar aws sam cli libexec lib site packages botocore client py line in make api call raise error class parsed response operation name botocore errorfactory nosuchbucket an error occurred nosuchbucket when calling the putobject operation the specified bucket does not exist the above exception was the direct cause of the following exception traceback most recent call last file opt homebrew cellar aws sam cli libexec lib site packages samcli lib package packageable resources py line in export self do export resource id resource dict parent dir file opt homebrew cellar aws sam cli libexec lib site packages samcli lib package packageable resources py line in do export uploaded url upload local artifacts file opt homebrew cellar aws sam cli libexec lib site packages samcli lib package utils py line in upload local artifacts return zip and upload local path uploader extension file opt homebrew cellar aws sam cli libexec lib site packages samcli lib package utils py line in zip and upload return uploader upload with dedup zip file precomputed hash extension extension file opt homebrew cellar aws sam cli libexec lib site packages samcli lib package uploader py line in upload with dedup return self upload file name remote path file opt homebrew cellar aws sam cli libexec lib site packages samcli lib package uploader py line in upload raise nosuchbucketerror bucket name self bucket name from ex samcli commands package exceptions nosuchbucketerror bucket does not exist sending telemetry metrics telemetry response error unable to upload artifact helloworldfunction referenced by codeuri parameter of helloworldfunction resource bucket does not exist expected result i would expect a sync to honor the settings in samconfig toml or b a cli flag to set the bucket name additional environment details ex windows mac amazon linux etc os mac os monterey if using sam cli sam version sam cli version aws region us west
0
8,043
2,611,449,702
IssuesEvent
2015-02-27 04:58:19
chrsmith/hedgewars
https://api.github.com/repos/chrsmith/hedgewars
closed
Missing tombstones prevent game launch
auto-migrated Engine Priority-Medium Type-Defect
``` (clone of http://fireforge.net//tracker/?func=detail&atid=125&aid=274&group_id=11 ) Submitted by: Christopher Vagnetoft (noccy80) "Detailed description Several times tonite a multiplayer game has failed to load since tombstones haven't been found on one or more of the players' computers. A missing tombstone should not cause the load to fail, but rather revert to a default." ``` Original issue reported on code.google.com by `sheepyluva` on 28 Jun 2010 at 8:28
1.0
Missing tombstones prevent game launch - ``` (clone of http://fireforge.net//tracker/?func=detail&atid=125&aid=274&group_id=11 ) Submitted by: Christopher Vagnetoft (noccy80) "Detailed description Several times tonite a multiplayer game has failed to load since tombstones haven't been found on one or more of the players' computers. A missing tombstone should not cause the load to fail, but rather revert to a default." ``` Original issue reported on code.google.com by `sheepyluva` on 28 Jun 2010 at 8:28
defect
missing tombstones prevent game launch clone of submitted by christopher vagnetoft detailed description several times tonite a multiplayer game has failed to load since tombstones haven t been found on one or more of the players computers a missing tombstone should not cause the load to fail but rather revert to a default original issue reported on code google com by sheepyluva on jun at
1
43,514
23,270,872,739
IssuesEvent
2022-08-04 22:55:03
quick-lint/quick-lint-js
https://api.github.com/repos/quick-lint/quick-lint-js
closed
Optimize Windows .exe icon
performance
dist/artwork/dusty-app.ico is 260 KiB. That's pretty big. Let's shrink it to bring the .exe file size down.
True
Optimize Windows .exe icon - dist/artwork/dusty-app.ico is 260 KiB. That's pretty big. Let's shrink it to bring the .exe file size down.
non_defect
optimize windows exe icon dist artwork dusty app ico is kib that s pretty big let s shrink it to bring the exe file size down
0
99,881
8,714,077,373
IssuesEvent
2018-12-07 06:13:37
actiontech/dble
https://api.github.com/repos/actiontech/dble
closed
System parameter `maxCon` is invalid after a connection failure
from auto_test resolve verified
* **dble version:** 5.6.29-dble-9.9.9.9-c53c3f7-20181116022818 * **preconditions :** no * **configs:** **schema.xml** ``` <?xml version='1.0' encoding='utf-8'?> <!DOCTYPE dble:schema SYSTEM "schema.dtd"><dble:schema xmlns:dble="http://dble.cloud/"> <schema dataNode="dn5" name="mytest" sqlMaxLimit="100"> <table dataNode="dn1,dn2,dn3,dn4" name="test" type="global" /> </schema> <dataNode dataHost="172.100.9.5" database="db1" name="dn1" /> <dataNode dataHost="172.100.9.6" database="db1" name="dn2" /> <dataNode dataHost="172.100.9.5" database="db2" name="dn3" /> <dataNode dataHost="172.100.9.6" database="db2" name="dn4" /> <dataNode dataHost="172.100.9.5" database="db3" name="dn5" /> <dataHost balance="0" maxCon="1000" minCon="10" name="172.100.9.5" slaveThreshold="100" switchType="1"> <heartbeat>select user()</heartbeat> <writeHost host="hostM1" password="111111" url="172.100.9.5:3306" user="test"> </writeHost> </dataHost> <dataHost balance="0" maxCon="1000" minCon="10" name="172.100.9.6" slaveThreshold="100" switchType="1"> <heartbeat>select user()</heartbeat> <writeHost host="hostM2" password="111111" url="172.100.9.6:3306" user="test"> </writeHost> </dataHost> </dble:schema> ``` **rule.xml** ``` <?xml version='1.0' encoding='utf-8'?> <!DOCTYPE dble:rule SYSTEM "rule.dtd"><dble:rule xmlns:dble="http://dble.cloud/"> <tableRule name="hash-two"> <rule> <columns>id</columns> <algorithm>two-long</algorithm> </rule> </tableRule> <tableRule name="hash-three"> <rule> <columns>id</columns> <algorithm>three-long</algorithm> </rule> </tableRule> <tableRule name="hash-four"> <rule> <columns>id</columns> <algorithm>four-long</algorithm> </rule> </tableRule> <function class="Hash" name="two-long"> <property name="partitionCount">2</property> <property name="partitionLength">1</property> </function> <function class="Hash" name="three-long"> <property name="partitionCount">3</property> <property name="partitionLength">1</property> </function> <function class="Hash" name="four-long"> <property name="partitionCount">4</property> <property name="partitionLength">1</property> </function> </dble:rule> ``` **server.xml** ``` <?xml version='1.0' encoding='utf-8'?> <!DOCTYPE dble:server SYSTEM "server.dtd"> <dble:server xmlns:dble="http://dble.cloud/"> <system> <property name="defaultSqlParser">druidparser</property> <property name="useGlobleTableCheck">1</property> <property name="processors">1</property> <property name="processorExecutor">1</property> <property name="maxCon">1</property> </system> <user name="root"> <property name="password">111111</property> <property name="manager">true</property> </user> <user name="test"> <property name="password">test</property> <property name="schemas">mytest</property> <property name="maxCon">1</property> </user> <user name="action"> <property name="password">action</property> <property name="schemas">mytest</property> <property name="readOnly">true</property> <property name="maxCon">1</property> </user> ``` * **steps:** step1. Do not close the connection after connecting to the client using the test user ``` [root@10-186-24-44 dble]# mysql -utest -ptest -h127.0.0.1 -P8066 mysql: [Warning] Using a password on the command line interface can be insecure. Welcome to the MySQL monitor. Commands end with ; or \g. Your MySQL connection id is 4 Server version: 5.6.29-dble-9.9.9.9-c53c3f7-20181116022818 dble Server (ActionTech) Copyright (c) 2000, 2016, Oracle and/or its affiliates. All rights reserved. Oracle is a registered trademark of Oracle Corporation and/or its affiliates. Other names may be trademarks of their respective owners. Type 'help;' or '\h' for help. Type '\c' to clear the current input statement. mysql> ``` step2. connect to the client using the test user ``` [root@10-186-24-44 dble]# mysql -utest -ptest -h127.0.0.1 -P8066 mysql: [Warning] Using a password on the command line interface can be insecure. ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111) ``` step3. connect to the client using the action user ``` [root@10-186-24-44 logs]# mysql -uaction -paction -h127.0.0.1 -P8066 mysql: [Warning] Using a password on the command line interface can be insecure. Welcome to the MySQL monitor. Commands end with ; or \g. Your MySQL connection id is 3 Server version: 5.6.29-dble-9.9.9.9-c53c3f7-20181116022818 dble Server (ActionTech) Copyright (c) 2000, 2016, Oracle and/or its affiliates. All rights reserved. Oracle is a registered trademark of Oracle Corporation and/or its affiliates. Other names may be trademarks of their respective owners. Type 'help;' or '\h' for help. Type '\c' to clear the current input statement. mysql> ``` * **expect result:** 1. step3: Connection failed * **real result:** 1. step3: Connection successed * **supplements:** 1. /label ~BUG
1.0
System parameter `maxCon` is invalid after a connection failure - * **dble version:** 5.6.29-dble-9.9.9.9-c53c3f7-20181116022818 * **preconditions :** no * **configs:** **schema.xml** ``` <?xml version='1.0' encoding='utf-8'?> <!DOCTYPE dble:schema SYSTEM "schema.dtd"><dble:schema xmlns:dble="http://dble.cloud/"> <schema dataNode="dn5" name="mytest" sqlMaxLimit="100"> <table dataNode="dn1,dn2,dn3,dn4" name="test" type="global" /> </schema> <dataNode dataHost="172.100.9.5" database="db1" name="dn1" /> <dataNode dataHost="172.100.9.6" database="db1" name="dn2" /> <dataNode dataHost="172.100.9.5" database="db2" name="dn3" /> <dataNode dataHost="172.100.9.6" database="db2" name="dn4" /> <dataNode dataHost="172.100.9.5" database="db3" name="dn5" /> <dataHost balance="0" maxCon="1000" minCon="10" name="172.100.9.5" slaveThreshold="100" switchType="1"> <heartbeat>select user()</heartbeat> <writeHost host="hostM1" password="111111" url="172.100.9.5:3306" user="test"> </writeHost> </dataHost> <dataHost balance="0" maxCon="1000" minCon="10" name="172.100.9.6" slaveThreshold="100" switchType="1"> <heartbeat>select user()</heartbeat> <writeHost host="hostM2" password="111111" url="172.100.9.6:3306" user="test"> </writeHost> </dataHost> </dble:schema> ``` **rule.xml** ``` <?xml version='1.0' encoding='utf-8'?> <!DOCTYPE dble:rule SYSTEM "rule.dtd"><dble:rule xmlns:dble="http://dble.cloud/"> <tableRule name="hash-two"> <rule> <columns>id</columns> <algorithm>two-long</algorithm> </rule> </tableRule> <tableRule name="hash-three"> <rule> <columns>id</columns> <algorithm>three-long</algorithm> </rule> </tableRule> <tableRule name="hash-four"> <rule> <columns>id</columns> <algorithm>four-long</algorithm> </rule> </tableRule> <function class="Hash" name="two-long"> <property name="partitionCount">2</property> <property name="partitionLength">1</property> </function> <function class="Hash" name="three-long"> <property name="partitionCount">3</property> <property name="partitionLength">1</property> </function> <function class="Hash" name="four-long"> <property name="partitionCount">4</property> <property name="partitionLength">1</property> </function> </dble:rule> ``` **server.xml** ``` <?xml version='1.0' encoding='utf-8'?> <!DOCTYPE dble:server SYSTEM "server.dtd"> <dble:server xmlns:dble="http://dble.cloud/"> <system> <property name="defaultSqlParser">druidparser</property> <property name="useGlobleTableCheck">1</property> <property name="processors">1</property> <property name="processorExecutor">1</property> <property name="maxCon">1</property> </system> <user name="root"> <property name="password">111111</property> <property name="manager">true</property> </user> <user name="test"> <property name="password">test</property> <property name="schemas">mytest</property> <property name="maxCon">1</property> </user> <user name="action"> <property name="password">action</property> <property name="schemas">mytest</property> <property name="readOnly">true</property> <property name="maxCon">1</property> </user> ``` * **steps:** step1. Do not close the connection after connecting to the client using the test user ``` [root@10-186-24-44 dble]# mysql -utest -ptest -h127.0.0.1 -P8066 mysql: [Warning] Using a password on the command line interface can be insecure. Welcome to the MySQL monitor. Commands end with ; or \g. Your MySQL connection id is 4 Server version: 5.6.29-dble-9.9.9.9-c53c3f7-20181116022818 dble Server (ActionTech) Copyright (c) 2000, 2016, Oracle and/or its affiliates. All rights reserved. Oracle is a registered trademark of Oracle Corporation and/or its affiliates. Other names may be trademarks of their respective owners. Type 'help;' or '\h' for help. Type '\c' to clear the current input statement. mysql> ``` step2. connect to the client using the test user ``` [root@10-186-24-44 dble]# mysql -utest -ptest -h127.0.0.1 -P8066 mysql: [Warning] Using a password on the command line interface can be insecure. ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111) ``` step3. connect to the client using the action user ``` [root@10-186-24-44 logs]# mysql -uaction -paction -h127.0.0.1 -P8066 mysql: [Warning] Using a password on the command line interface can be insecure. Welcome to the MySQL monitor. Commands end with ; or \g. Your MySQL connection id is 3 Server version: 5.6.29-dble-9.9.9.9-c53c3f7-20181116022818 dble Server (ActionTech) Copyright (c) 2000, 2016, Oracle and/or its affiliates. All rights reserved. Oracle is a registered trademark of Oracle Corporation and/or its affiliates. Other names may be trademarks of their respective owners. Type 'help;' or '\h' for help. Type '\c' to clear the current input statement. mysql> ``` * **expect result:** 1. step3: Connection failed * **real result:** 1. step3: Connection successed * **supplements:** 1. /label ~BUG
non_defect
system parameter maxcon is invalid after a connection failure dble version: dble preconditions : no configs: schema xml dble schema xmlns dble select user select user rule xml dble rule xmlns dble id two long id three long id four long server xml dble server xmlns dble druidparser true test mytest action mytest true steps: do not close the connection after connecting to the client using the test user mysql utest ptest mysql using a password on the command line interface can be insecure welcome to the mysql monitor commands end with or g your mysql connection id is server version dble dble server actiontech copyright c oracle and or its affiliates all rights reserved oracle is a registered trademark of oracle corporation and or its affiliates other names may be trademarks of their respective owners type help or h for help type c to clear the current input statement mysql connect to the client using the test user mysql utest ptest mysql using a password on the command line interface can be insecure error can t connect to mysql server on connect to the client using the action user mysql uaction paction mysql using a password on the command line interface can be insecure welcome to the mysql monitor commands end with or g your mysql connection id is server version dble dble server actiontech copyright c oracle and or its affiliates all rights reserved oracle is a registered trademark of oracle corporation and or its affiliates other names may be trademarks of their respective owners type help or h for help type c to clear the current input statement mysql expect result: connection failed real result: connection successed supplements: label bug
0
16,097
2,871,807,576
IssuesEvent
2015-06-08 07:34:08
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
reopened
[TEST-FAILURE] MigrationAwareServiceTest.testPartitionDataSize_whenNodesStartedParallel_withSingleBackup
Team: Core Type: Defect
``` java.lang.AssertionError: expected:<542> but was:<506> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:834) at org.junit.Assert.assertEquals(Assert.java:645) ``` https://hazelcast-l337.ci.cloudbees.com/job/Hazelcast-3.x-OpenJDK8-Quality-Outreach/com.hazelcast$hazelcast/72/testReport/junit/com.hazelcast.spi/MigrationAwareServiceTest/testPartitionDataSize_whenNodesStartedParallel_withSingleBackup/
1.0
[TEST-FAILURE] MigrationAwareServiceTest.testPartitionDataSize_whenNodesStartedParallel_withSingleBackup - ``` java.lang.AssertionError: expected:<542> but was:<506> at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:834) at org.junit.Assert.assertEquals(Assert.java:645) ``` https://hazelcast-l337.ci.cloudbees.com/job/Hazelcast-3.x-OpenJDK8-Quality-Outreach/com.hazelcast$hazelcast/72/testReport/junit/com.hazelcast.spi/MigrationAwareServiceTest/testPartitionDataSize_whenNodesStartedParallel_withSingleBackup/
defect
migrationawareservicetest testpartitiondatasize whennodesstartedparallel withsinglebackup java lang assertionerror expected but was at org junit assert fail assert java at org junit assert failnotequals assert java at org junit assert assertequals assert java
1
54,837
13,960,445,927
IssuesEvent
2020-10-24 21:07:12
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
opened
'configure' fails on Kernel 5.8 with no module support
Status: Triage Needed Type: Defect
### System information Type | Version/Name --- | --- Distribution Name | Gentoo Linux Kernel | 5.8+ Architecture | amd64 ZFS Version | 0.8.6-staging, 2.0.0-rc4 ### Describe the problem you're observing Configure fails when used in builtin mode for Linux 5.8 and 5.9 when those have no support for loadable modules. It works on Linux 5.7 and lower with the same ZFS versions mentioned earlier. The command I use: ```console $ ./configure --with-linux=/path/to/linux-stable.git --enable-linux-builtin [...] checking kernel source directory... /path/to/linux-stable.git checking kernel build directory... /path/to/linux-stable.git checking kernel source version... 5.8.16 checking kernel file name for module symbols... NONE checking whether modules can be built... no configure: error: *** Unable to build an empty module. *** Please run 'make scripts' inside the kernel source tree. ``` ### Include any warning/errors/backtraces from the system logs Here is the extract of `config.log`: ``` configure:49874: checking whether modules can be built configure:49935: KBUILD_MODPOST_NOFINAL= KBUILD_MODPOST_WARN= make modules -k -j8 -C /path/to/linux-stable.git M=/path/to/zfs.git/build/conftest >build/conftest/build.log 2>&1 configure:49938: $? = 2 configure:49941: test -f build/conftest/conftest.o configure:49944: $? = 1 configure:49952: result: no configure:49959: error: *** Unable to build an empty module. *** Please run 'make scripts' inside the kernel source tree. ``` More interestingly, here is the content of `build/conftest/build.log`: ``` make: Entering directory '/path/to/linux-stable.git' make[1]: *** No rule to make target '/path/to/zfs.git/build/conftest/modules.order', needed by '/path/to/zfs.git/build/conftest/Module.symvers'. WARNING: Symbol version dump "Module.symvers" is missing. Modules may not have dependencies or modversions. make[1]: Target '__modpost' not remade because of errors. make: *** [Makefile:1665: modules] Error 2 make: Leaving directory '/path/to/linux-stable.git' ``` ### Further analysis in kernel changes Of course this sudden breakage cannot be attributed to ZFS but this has a direct impact when you want to compile ZFS in a standalone kernel. I did a little bisect in the kernel sources between 5.7 and 5.8 and found that the following commit is responsible: - https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/commit/?id=436b2ac603d58504f38041a0cd8adb5aeace992b It seems to be part of a bigger refactoring around modpost performed by Masahiro Yamada (@masahir0y). There may be a workaround but honestly I don't fully understand what really causes this issue.
1.0
'configure' fails on Kernel 5.8 with no module support - ### System information Type | Version/Name --- | --- Distribution Name | Gentoo Linux Kernel | 5.8+ Architecture | amd64 ZFS Version | 0.8.6-staging, 2.0.0-rc4 ### Describe the problem you're observing Configure fails when used in builtin mode for Linux 5.8 and 5.9 when those have no support for loadable modules. It works on Linux 5.7 and lower with the same ZFS versions mentioned earlier. The command I use: ```console $ ./configure --with-linux=/path/to/linux-stable.git --enable-linux-builtin [...] checking kernel source directory... /path/to/linux-stable.git checking kernel build directory... /path/to/linux-stable.git checking kernel source version... 5.8.16 checking kernel file name for module symbols... NONE checking whether modules can be built... no configure: error: *** Unable to build an empty module. *** Please run 'make scripts' inside the kernel source tree. ``` ### Include any warning/errors/backtraces from the system logs Here is the extract of `config.log`: ``` configure:49874: checking whether modules can be built configure:49935: KBUILD_MODPOST_NOFINAL= KBUILD_MODPOST_WARN= make modules -k -j8 -C /path/to/linux-stable.git M=/path/to/zfs.git/build/conftest >build/conftest/build.log 2>&1 configure:49938: $? = 2 configure:49941: test -f build/conftest/conftest.o configure:49944: $? = 1 configure:49952: result: no configure:49959: error: *** Unable to build an empty module. *** Please run 'make scripts' inside the kernel source tree. ``` More interestingly, here is the content of `build/conftest/build.log`: ``` make: Entering directory '/path/to/linux-stable.git' make[1]: *** No rule to make target '/path/to/zfs.git/build/conftest/modules.order', needed by '/path/to/zfs.git/build/conftest/Module.symvers'. WARNING: Symbol version dump "Module.symvers" is missing. Modules may not have dependencies or modversions. make[1]: Target '__modpost' not remade because of errors. make: *** [Makefile:1665: modules] Error 2 make: Leaving directory '/path/to/linux-stable.git' ``` ### Further analysis in kernel changes Of course this sudden breakage cannot be attributed to ZFS but this has a direct impact when you want to compile ZFS in a standalone kernel. I did a little bisect in the kernel sources between 5.7 and 5.8 and found that the following commit is responsible: - https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/commit/?id=436b2ac603d58504f38041a0cd8adb5aeace992b It seems to be part of a bigger refactoring around modpost performed by Masahiro Yamada (@masahir0y). There may be a workaround but honestly I don't fully understand what really causes this issue.
defect
configure fails on kernel with no module support system information type version name distribution name gentoo linux kernel architecture zfs version staging describe the problem you re observing configure fails when used in builtin mode for linux and when those have no support for loadable modules it works on linux and lower with the same zfs versions mentioned earlier the command i use console configure with linux path to linux stable git enable linux builtin checking kernel source directory path to linux stable git checking kernel build directory path to linux stable git checking kernel source version checking kernel file name for module symbols none checking whether modules can be built no configure error unable to build an empty module please run make scripts inside the kernel source tree include any warning errors backtraces from the system logs here is the extract of config log configure checking whether modules can be built configure kbuild modpost nofinal kbuild modpost warn make modules k c path to linux stable git m path to zfs git build conftest build conftest build log configure configure test f build conftest conftest o configure configure result no configure error unable to build an empty module please run make scripts inside the kernel source tree more interestingly here is the content of build conftest build log make entering directory path to linux stable git make no rule to make target path to zfs git build conftest modules order needed by path to zfs git build conftest module symvers warning symbol version dump module symvers is missing modules may not have dependencies or modversions make target modpost not remade because of errors make error make leaving directory path to linux stable git further analysis in kernel changes of course this sudden breakage cannot be attributed to zfs but this has a direct impact when you want to compile zfs in a standalone kernel i did a little bisect in the kernel sources between and and found that the following commit is responsible it seems to be part of a bigger refactoring around modpost performed by masahiro yamada there may be a workaround but honestly i don t fully understand what really causes this issue
1
18,541
2,615,173,145
IssuesEvent
2015-03-01 06:55:20
chrsmith/html5rocks
https://api.github.com/repos/chrsmith/html5rocks
closed
typo in the fieldguide
auto-migrated fieldguide Milestone-Q12012 Priority-P3 Type-Bug
``` "A web application providers a great user experience" --------------------------^ at http://www.html5rocks.com/webappfieldguide/know-your-apps/site-vs-app/ ``` Original issue reported on code.google.com by `mikenere...@gmail.com` on 15 Feb 2012 at 2:51
1.0
typo in the fieldguide - ``` "A web application providers a great user experience" --------------------------^ at http://www.html5rocks.com/webappfieldguide/know-your-apps/site-vs-app/ ``` Original issue reported on code.google.com by `mikenere...@gmail.com` on 15 Feb 2012 at 2:51
non_defect
typo in the fieldguide a web application providers a great user experience at original issue reported on code google com by mikenere gmail com on feb at
0
217,939
16,891,531,463
IssuesEvent
2021-06-23 09:49:48
hakehuang/zephyr
https://api.github.com/repos/hakehuang/zephyr
opened
tests-ci :kernel.common.stack_protection_no_userspace.fatal : zephyr-v2.6.0-286-g46029914a7ac: mimxrt1015_evk: test Flash error
area: Tests bug
**Describe the bug** kernel.common.stack_protection_no_userspace.fatal test is Flash error on zephyr-v2.6.0-286-g46029914a7ac on mimxrt1015_evk see logs for details **To Reproduce** 1. ``` scripts/twister --device-testing --device-serial /dev/ttyACM0 -p mimxrt1015_evk --testcase-root tests --sub-test kernel.common ``` 2. See error **Expected behavior** test pass **Impact** **Logs and console output** ``` None ``` **Environment (please complete the following information):** - OS: (e.g. Linux ) - Toolchain (e.g Zephyr SDK) - Commit SHA or Version used: zephyr-v2.6.0-286-g46029914a7ac
1.0
tests-ci :kernel.common.stack_protection_no_userspace.fatal : zephyr-v2.6.0-286-g46029914a7ac: mimxrt1015_evk: test Flash error - **Describe the bug** kernel.common.stack_protection_no_userspace.fatal test is Flash error on zephyr-v2.6.0-286-g46029914a7ac on mimxrt1015_evk see logs for details **To Reproduce** 1. ``` scripts/twister --device-testing --device-serial /dev/ttyACM0 -p mimxrt1015_evk --testcase-root tests --sub-test kernel.common ``` 2. See error **Expected behavior** test pass **Impact** **Logs and console output** ``` None ``` **Environment (please complete the following information):** - OS: (e.g. Linux ) - Toolchain (e.g Zephyr SDK) - Commit SHA or Version used: zephyr-v2.6.0-286-g46029914a7ac
non_defect
tests ci kernel common stack protection no userspace fatal zephyr evk test flash error describe the bug kernel common stack protection no userspace fatal test is flash error on zephyr on evk see logs for details to reproduce scripts twister device testing device serial dev p evk testcase root tests sub test kernel common see error expected behavior test pass impact logs and console output none environment please complete the following information os e g linux toolchain e g zephyr sdk commit sha or version used zephyr
0
398,611
11,741,997,173
IssuesEvent
2020-03-11 23:16:32
thaliawww/concrexit
https://api.github.com/repos/thaliawww/concrexit
closed
Toon dagnaam in events admin
events priority: low
In GitLab by njanssen on Mar 1, 2017, 20:05 Het zou fijn zijn als de naam van de dag (e.g. maandag, dinsdag, ...) in de datumweergaven van de backend staat. Edit: Het gaat hierbij om de weergave van de evenementen.
1.0
Toon dagnaam in events admin - In GitLab by njanssen on Mar 1, 2017, 20:05 Het zou fijn zijn als de naam van de dag (e.g. maandag, dinsdag, ...) in de datumweergaven van de backend staat. Edit: Het gaat hierbij om de weergave van de evenementen.
non_defect
toon dagnaam in events admin in gitlab by njanssen on mar het zou fijn zijn als de naam van de dag e g maandag dinsdag in de datumweergaven van de backend staat edit het gaat hierbij om de weergave van de evenementen
0
2,245
2,712,129,537
IssuesEvent
2015-04-09 11:46:27
HGustavs/LenaSYS
https://api.github.com/repos/HGustavs/LenaSYS
closed
The Back/Return button
CodeViewer
The Return/Back button in codeviewer isn't linked to a page, when clicked on. <-a href="sectioned.php?courseid=UNK&amp;coursevers=UNK"><-img src="../Shared/icons/Up.svg"-><-/a> The linked page is missing in the server(?)
1.0
The Back/Return button - The Return/Back button in codeviewer isn't linked to a page, when clicked on. <-a href="sectioned.php?courseid=UNK&amp;coursevers=UNK"><-img src="../Shared/icons/Up.svg"-><-/a> The linked page is missing in the server(?)
non_defect
the back return button the return back button in codeviewer isn t linked to a page when clicked on the linked page is missing in the server
0
47,362
2,978,303,478
IssuesEvent
2015-07-16 04:47:27
pombase/canto
https://api.github.com/repos/pombase/canto
closed
Do we need the New curs form?
admin low priority
Can we remove the New curs page (reached via the "Add ... Curation session" admin link)? It looks buggy in the test instance right now (and I ain't about to mess with what's in live), and I don't think it would be the most useful thing even if it got fixed. Doesn't seem like fun to generate hex-gibberish session IDs manually. We can easily do without it, since we can find and add papers, and then push the button to create sessions for them, with autogenerated IDs creation dates.
1.0
Do we need the New curs form? - Can we remove the New curs page (reached via the "Add ... Curation session" admin link)? It looks buggy in the test instance right now (and I ain't about to mess with what's in live), and I don't think it would be the most useful thing even if it got fixed. Doesn't seem like fun to generate hex-gibberish session IDs manually. We can easily do without it, since we can find and add papers, and then push the button to create sessions for them, with autogenerated IDs creation dates.
non_defect
do we need the new curs form can we remove the new curs page reached via the add curation session admin link it looks buggy in the test instance right now and i ain t about to mess with what s in live and i don t think it would be the most useful thing even if it got fixed doesn t seem like fun to generate hex gibberish session ids manually we can easily do without it since we can find and add papers and then push the button to create sessions for them with autogenerated ids creation dates
0
32,386
6,767,416,567
IssuesEvent
2017-10-26 03:11:47
Shopkeepers/Shopkeepers
https://api.github.com/repos/Shopkeepers/Shopkeepers
closed
With-Shop-creation near spawn->Witch blocked->Server restart->witch spawns but not editable
Defect fixed migrated
**Migrated from:** https://dev.bukkit.org/projects/shopkeepers/issues/111 **Originally posted by blablubbabc (Mar 29, 2013):** I just discovered the following:- I have a zone in my world, where monster spawning is denied by a plugin. - I tried setting up a witch shop there: it displayed me the message: "blabla, shop created, edit it!" but the witch of course was not spawned - After a server restart (no reload), the witch is there.., BUT is not working ( I can't click it to open menu etc. .., but I still can't attack it..) - I took a look at the save.yml and saw some other witch-shops also being there (which I must have tried to place them earlier in the same zone), BUT these witches are not placed at the locations saved, like the one described above.. - the witch, which was placed and can't be removed from ingame, is near the spawn (spawn chunk).. maybe this matters..Maybe you should block shop creation if an other plugin is blocking something or override the cancling of other plugins ( would be nice to be able to have witch-shops, while other monsters are blocked from spawning..THEN: I just set debug to true in the config and reload the shopkeepers plugin via PluginManager (which seemed to work fine): The witch is gone, and in the console there are messages for each not spawned witch: "[WARNING] [Shopkeepers] Failed to spawn shopkeeper at world,434,71,722"Also: Did you know that traders placed in tall grass are constantly ported back every 5 seconds for "being pushed", because their location was saved to the block above the tall grass and they fall through this block by their own all the time.. Maybe you could check for blocks like this (maybe add a list of "passable) block-types to the config.. peopel can then add water ect. to it if they want..) and set their location to the lowest possible location, going down all "passable" blocks on creation
1.0
With-Shop-creation near spawn->Witch blocked->Server restart->witch spawns but not editable - **Migrated from:** https://dev.bukkit.org/projects/shopkeepers/issues/111 **Originally posted by blablubbabc (Mar 29, 2013):** I just discovered the following:- I have a zone in my world, where monster spawning is denied by a plugin. - I tried setting up a witch shop there: it displayed me the message: "blabla, shop created, edit it!" but the witch of course was not spawned - After a server restart (no reload), the witch is there.., BUT is not working ( I can't click it to open menu etc. .., but I still can't attack it..) - I took a look at the save.yml and saw some other witch-shops also being there (which I must have tried to place them earlier in the same zone), BUT these witches are not placed at the locations saved, like the one described above.. - the witch, which was placed and can't be removed from ingame, is near the spawn (spawn chunk).. maybe this matters..Maybe you should block shop creation if an other plugin is blocking something or override the cancling of other plugins ( would be nice to be able to have witch-shops, while other monsters are blocked from spawning..THEN: I just set debug to true in the config and reload the shopkeepers plugin via PluginManager (which seemed to work fine): The witch is gone, and in the console there are messages for each not spawned witch: "[WARNING] [Shopkeepers] Failed to spawn shopkeeper at world,434,71,722"Also: Did you know that traders placed in tall grass are constantly ported back every 5 seconds for "being pushed", because their location was saved to the block above the tall grass and they fall through this block by their own all the time.. Maybe you could check for blocks like this (maybe add a list of "passable) block-types to the config.. peopel can then add water ect. to it if they want..) and set their location to the lowest possible location, going down all "passable" blocks on creation
defect
with shop creation near spawn witch blocked server restart witch spawns but not editable migrated from originally posted by blablubbabc mar i just discovered the following i have a zone in my world where monster spawning is denied by a plugin i tried setting up a witch shop there it displayed me the message blabla shop created edit it but the witch of course was not spawned after a server restart no reload the witch is there but is not working i can t click it to open menu etc but i still can t attack it i took a look at the save yml and saw some other witch shops also being there which i must have tried to place them earlier in the same zone but these witches are not placed at the locations saved like the one described above the witch which was placed and can t be removed from ingame is near the spawn spawn chunk maybe this matters maybe you should block shop creation if an other plugin is blocking something or override the cancling of other plugins would be nice to be able to have witch shops while other monsters are blocked from spawning then i just set debug to true in the config and reload the shopkeepers plugin via pluginmanager which seemed to work fine the witch is gone and in the console there are messages for each not spawned witch failed to spawn shopkeeper at world also did you know that traders placed in tall grass are constantly ported back every seconds for being pushed because their location was saved to the block above the tall grass and they fall through this block by their own all the time maybe you could check for blocks like this maybe add a list of passable block types to the config peopel can then add water ect to it if they want and set their location to the lowest possible location going down all passable blocks on creation
1
45,111
11,589,787,925
IssuesEvent
2020-02-24 03:54:54
GoogleContainerTools/skaffold
https://api.github.com/repos/GoogleContainerTools/skaffold
reopened
Get build timeout on local concurrent build
area/build kind/bug
### Actual behavior with `skaffold dev` i rarely get this sort of error: ``` Sending build context to Docker daemon 9.742MB Step 1/10 : FROM node:10.15 FATA[1032] failed to build: build failed: building [us.gcr.io/replay-gaming/poker-api]: build artifact: unable to stream build output: Get https://registry-1.docker.io/v2/library/node/manifests/10.15: net/http: TLS handshake timeout make[1]: *** [dependencies] Error 1 make: *** [init] Error 2 ``` Keep in mind, that building of all images takes more than 15 minutes ### Information - Skaffold version: 1.3.1 - Operating system: MacOs 10.15 - Contents of skaffold.yaml: ```yaml apiVersion: skaffold/v2alpha3 kind: Config build: local: concurrency: 0 tagPolicy: sha256: {} artifacts: - image: us.gcr.io/replay-gaming/rails docker: dockerfile: docker/development/Dockerfile context: rails sync: infer: - 'app/**/*' - 'bin/**/*' - 'db/**/*' - 'lib/**/*' - 'public/**/*' - 'script/**/*' - 'spec/**/*' - image: us.gcr.io/replay-gaming/poker-core docker: dockerfile: docker/development/Dockerfile context: core sync: infer: - 'bin/**/*' - 'db/**/*' - 'lib/**/*' - 'system/**/*' - 'schemas/**/*' - 'spec/**/*' - image: us.gcr.io/replay-gaming/poker-api docker: dockerfile: docker/development/Dockerfile context: api sync: infer: - 'bin/**/*' - 'src/**/*' - 'tests/**/*' - image: us.gcr.io/replay-gaming/poker-client docker: dockerfile: docker/development/Dockerfile context: client sync: infer: - 'bin/**/*' - 'public/**/*' - 'src/**/*' - image: us.gcr.io/replay-gaming/poker-scheduler docker: dockerfile: docker/development/Dockerfile context: scheduler sync: infer: - 'src/**/*' - image: us.gcr.io/replay-gaming/go-eventsource context: es ``` I assume that the root of the problem is in `concurrency` setting. ### Steps to reproduce the behavior `skaffold dev` This is flaky!
1.0
Get build timeout on local concurrent build - ### Actual behavior with `skaffold dev` i rarely get this sort of error: ``` Sending build context to Docker daemon 9.742MB Step 1/10 : FROM node:10.15 FATA[1032] failed to build: build failed: building [us.gcr.io/replay-gaming/poker-api]: build artifact: unable to stream build output: Get https://registry-1.docker.io/v2/library/node/manifests/10.15: net/http: TLS handshake timeout make[1]: *** [dependencies] Error 1 make: *** [init] Error 2 ``` Keep in mind, that building of all images takes more than 15 minutes ### Information - Skaffold version: 1.3.1 - Operating system: MacOs 10.15 - Contents of skaffold.yaml: ```yaml apiVersion: skaffold/v2alpha3 kind: Config build: local: concurrency: 0 tagPolicy: sha256: {} artifacts: - image: us.gcr.io/replay-gaming/rails docker: dockerfile: docker/development/Dockerfile context: rails sync: infer: - 'app/**/*' - 'bin/**/*' - 'db/**/*' - 'lib/**/*' - 'public/**/*' - 'script/**/*' - 'spec/**/*' - image: us.gcr.io/replay-gaming/poker-core docker: dockerfile: docker/development/Dockerfile context: core sync: infer: - 'bin/**/*' - 'db/**/*' - 'lib/**/*' - 'system/**/*' - 'schemas/**/*' - 'spec/**/*' - image: us.gcr.io/replay-gaming/poker-api docker: dockerfile: docker/development/Dockerfile context: api sync: infer: - 'bin/**/*' - 'src/**/*' - 'tests/**/*' - image: us.gcr.io/replay-gaming/poker-client docker: dockerfile: docker/development/Dockerfile context: client sync: infer: - 'bin/**/*' - 'public/**/*' - 'src/**/*' - image: us.gcr.io/replay-gaming/poker-scheduler docker: dockerfile: docker/development/Dockerfile context: scheduler sync: infer: - 'src/**/*' - image: us.gcr.io/replay-gaming/go-eventsource context: es ``` I assume that the root of the problem is in `concurrency` setting. ### Steps to reproduce the behavior `skaffold dev` This is flaky!
non_defect
get build timeout on local concurrent build actual behavior with skaffold dev i rarely get this sort of error sending build context to docker daemon step from node fata failed to build build failed building build artifact unable to stream build output get net http tls handshake timeout make error make error keep in mind that building of all images takes more than minutes information skaffold version operating system macos contents of skaffold yaml yaml apiversion skaffold kind config build local concurrency tagpolicy artifacts image us gcr io replay gaming rails docker dockerfile docker development dockerfile context rails sync infer app bin db lib public script spec image us gcr io replay gaming poker core docker dockerfile docker development dockerfile context core sync infer bin db lib system schemas spec image us gcr io replay gaming poker api docker dockerfile docker development dockerfile context api sync infer bin src tests image us gcr io replay gaming poker client docker dockerfile docker development dockerfile context client sync infer bin public src image us gcr io replay gaming poker scheduler docker dockerfile docker development dockerfile context scheduler sync infer src image us gcr io replay gaming go eventsource context es i assume that the root of the problem is in concurrency setting steps to reproduce the behavior skaffold dev this is flaky
0
98,444
29,870,547,119
IssuesEvent
2023-06-20 08:12:16
GSS-Cogs/dd-cms
https://api.github.com/repos/GSS-Cogs/dd-cms
closed
Automate legend font change
chart builder high priority
At the moment the only way to update the legend font to the new font and size is to resave each chart.
1.0
Automate legend font change - At the moment the only way to update the legend font to the new font and size is to resave each chart.
non_defect
automate legend font change at the moment the only way to update the legend font to the new font and size is to resave each chart
0
56,792
15,371,686,401
IssuesEvent
2021-03-02 10:18:47
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
com.hazelcast.json.MapPredicateJsonTest takes 20+ minutes to complete
Module: IMap Module: Query Source: Internal Team: Core Type: Defect Type: Test-Failure
Tests in the `com.hazelcast.json.MapPredicateJsonTest` take 20+ minutes to complete in the PR builder. It might make sense to mark this as `SlowTest` but I feel like the tests there are pretty simple and should complete quickly. When I run the test suite locally, I see that in some tests, I get the following logs regarding invocation retries and individual tests take 1+ min to complete. Sometimes, the same test takes 2-3 seconds to complete. For example, for the `testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn` ``` 17:13:40,188 INFO |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [LifecycleService] Thread-9 - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] [127.0.0.1]:5703 is STARTED 17:13:40,189 INFO |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [PartitionStateManager] hz.condescending_jennings.priority-generic-operation.thread-0 - [127.0.0.1]:5701 [dev] [4.2-SNAPSHOT] Initializing cluster partition table arrangement... 17:14:14,766 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=114, invocationTime=1611324854764 (2021-01-22 17:14:14.764), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=100, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService 17:14:15,280 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=124, invocationTime=1611324855279 (2021-01-22 17:14:15.279), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=110, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService 17:14:20,284 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=134, invocationTime=1611324860283 (2021-01-22 17:14:20.283), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=120, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService 17:14:25,288 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=144, invocationTime=1611324865286 (2021-01-22 17:14:25.286), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=130, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService 17:14:26,418 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=154, invocationTime=1611324866417 (2021-01-22 17:14:26.417), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=140, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService 17:14:30,806 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=164, invocationTime=1611324870805 (2021-01-22 17:14:30.805), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=150, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService 17:14:35,811 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=174, invocationTime=1611324875809 (2021-01-22 17:14:35.809), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=160, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService 17:14:37,846 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=184, invocationTime=1611324877845 (2021-01-22 17:14:37.845), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=170, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService 17:14:41,331 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=194, invocationTime=1611324881329 (2021-01-22 17:14:41.329), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=180, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService ``` It would be good to investigate the root cause of this.
1.0
com.hazelcast.json.MapPredicateJsonTest takes 20+ minutes to complete - Tests in the `com.hazelcast.json.MapPredicateJsonTest` take 20+ minutes to complete in the PR builder. It might make sense to mark this as `SlowTest` but I feel like the tests there are pretty simple and should complete quickly. When I run the test suite locally, I see that in some tests, I get the following logs regarding invocation retries and individual tests take 1+ min to complete. Sometimes, the same test takes 2-3 seconds to complete. For example, for the `testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn` ``` 17:13:40,188 INFO |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [LifecycleService] Thread-9 - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] [127.0.0.1]:5703 is STARTED 17:13:40,189 INFO |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [PartitionStateManager] hz.condescending_jennings.priority-generic-operation.thread-0 - [127.0.0.1]:5701 [dev] [4.2-SNAPSHOT] Initializing cluster partition table arrangement... 17:14:14,766 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=114, invocationTime=1611324854764 (2021-01-22 17:14:14.764), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=100, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService 17:14:15,280 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=124, invocationTime=1611324855279 (2021-01-22 17:14:15.279), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=110, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService 17:14:20,284 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=134, invocationTime=1611324860283 (2021-01-22 17:14:20.283), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=120, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService 17:14:25,288 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=144, invocationTime=1611324865286 (2021-01-22 17:14:25.286), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=130, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService 17:14:26,418 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=154, invocationTime=1611324866417 (2021-01-22 17:14:26.417), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=140, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService 17:14:30,806 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=164, invocationTime=1611324870805 (2021-01-22 17:14:30.805), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=150, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService 17:14:35,811 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=174, invocationTime=1611324875809 (2021-01-22 17:14:35.809), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=160, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService 17:14:37,846 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=184, invocationTime=1611324877845 (2021-01-22 17:14:37.845), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=170, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService 17:14:41,331 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=194, invocationTime=1611324881329 (2021-01-22 17:14:41.329), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=180, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService ``` It would be good to investigate the root cause of this.
defect
com hazelcast json mappredicatejsontest takes minutes to complete tests in the com hazelcast json mappredicatejsontest take minutes to complete in the pr builder it might make sense to mark this as slowtest but i feel like the tests there are pretty simple and should complete quickly when i run the test suite locally i see that in some tests i get the following logs regarding invocation retries and individual tests take min to complete sometimes the same test takes seconds to complete for example for the testnonterminalattributeis queriedwithequalsnull shouldnotreturn info testnonterminalattributeis queriedwithequalsnull shouldnotreturn thread is started info testnonterminalattributeis queriedwithequalsnull shouldnotreturn hz condescending jennings priority generic operation thread initializing cluster partition table arrangement warn testnonterminalattributeis queriedwithequalsnull shouldnotreturn hz vigorous jennings invocationmonitorthread retrying invocation invocation op com hazelcast map impl query queryoperation servicename hz impl mapservice identityhash partitionid replicaindex callid invocationtime waittimeout calltimeout tenantcontrol com hazelcast spi impl tenantcontrol nooptenantcontrol name trycount trypausemillis invokecount calltimeoutmillis firstinvocationtimems firstinvocationtime lastheartbeatmillis lastheartbeattime target null pendingresponse void backupsacksexpected backupsacksreceived connection null reason com hazelcast spi exception wrongtargetexception wrongtarget local member this expected target null partitionid replicaindex operation com hazelcast map impl query queryoperation service hz impl mapservice warn testnonterminalattributeis queriedwithequalsnull shouldnotreturn hz vigorous jennings invocationmonitorthread retrying invocation invocation op com hazelcast map impl query queryoperation servicename hz impl mapservice identityhash partitionid replicaindex callid invocationtime waittimeout calltimeout tenantcontrol com hazelcast spi impl tenantcontrol nooptenantcontrol name trycount trypausemillis invokecount calltimeoutmillis firstinvocationtimems firstinvocationtime lastheartbeatmillis lastheartbeattime target null pendingresponse void backupsacksexpected backupsacksreceived connection null reason com hazelcast spi exception wrongtargetexception wrongtarget local member this expected target null partitionid replicaindex operation com hazelcast map impl query queryoperation service hz impl mapservice warn testnonterminalattributeis queriedwithequalsnull shouldnotreturn hz vigorous jennings invocationmonitorthread retrying invocation invocation op com hazelcast map impl query queryoperation servicename hz impl mapservice identityhash partitionid replicaindex callid invocationtime waittimeout calltimeout tenantcontrol com hazelcast spi impl tenantcontrol nooptenantcontrol name trycount trypausemillis invokecount calltimeoutmillis firstinvocationtimems firstinvocationtime lastheartbeatmillis lastheartbeattime target null pendingresponse void backupsacksexpected backupsacksreceived connection null reason com hazelcast spi exception wrongtargetexception wrongtarget local member this expected target null partitionid replicaindex operation com hazelcast map impl query queryoperation service hz impl mapservice warn testnonterminalattributeis queriedwithequalsnull shouldnotreturn hz vigorous jennings invocationmonitorthread retrying invocation invocation op com hazelcast map impl query queryoperation servicename hz impl mapservice identityhash partitionid replicaindex callid invocationtime waittimeout calltimeout tenantcontrol com hazelcast spi impl tenantcontrol nooptenantcontrol name trycount trypausemillis invokecount calltimeoutmillis firstinvocationtimems firstinvocationtime lastheartbeatmillis lastheartbeattime target null pendingresponse void backupsacksexpected backupsacksreceived connection null reason com hazelcast spi exception wrongtargetexception wrongtarget local member this expected target null partitionid replicaindex operation com hazelcast map impl query queryoperation service hz impl mapservice warn testnonterminalattributeis queriedwithequalsnull shouldnotreturn hz vigorous jennings invocationmonitorthread retrying invocation invocation op com hazelcast map impl query queryoperation servicename hz impl mapservice identityhash partitionid replicaindex callid invocationtime waittimeout calltimeout tenantcontrol com hazelcast spi impl tenantcontrol nooptenantcontrol name trycount trypausemillis invokecount calltimeoutmillis firstinvocationtimems firstinvocationtime lastheartbeatmillis lastheartbeattime target null pendingresponse void backupsacksexpected backupsacksreceived connection null reason com hazelcast spi exception wrongtargetexception wrongtarget local member this expected target null partitionid replicaindex operation com hazelcast map impl query queryoperation service hz impl mapservice warn testnonterminalattributeis queriedwithequalsnull shouldnotreturn hz vigorous jennings invocationmonitorthread retrying invocation invocation op com hazelcast map impl query queryoperation servicename hz impl mapservice identityhash partitionid replicaindex callid invocationtime waittimeout calltimeout tenantcontrol com hazelcast spi impl tenantcontrol nooptenantcontrol name trycount trypausemillis invokecount calltimeoutmillis firstinvocationtimems firstinvocationtime lastheartbeatmillis lastheartbeattime target null pendingresponse void backupsacksexpected backupsacksreceived connection null reason com hazelcast spi exception wrongtargetexception wrongtarget local member this expected target null partitionid replicaindex operation com hazelcast map impl query queryoperation service hz impl mapservice warn testnonterminalattributeis queriedwithequalsnull shouldnotreturn hz vigorous jennings invocationmonitorthread retrying invocation invocation op com hazelcast map impl query queryoperation servicename hz impl mapservice identityhash partitionid replicaindex callid invocationtime waittimeout calltimeout tenantcontrol com hazelcast spi impl tenantcontrol nooptenantcontrol name trycount trypausemillis invokecount calltimeoutmillis firstinvocationtimems firstinvocationtime lastheartbeatmillis lastheartbeattime target null pendingresponse void backupsacksexpected backupsacksreceived connection null reason com hazelcast spi exception wrongtargetexception wrongtarget local member this expected target null partitionid replicaindex operation com hazelcast map impl query queryoperation service hz impl mapservice warn testnonterminalattributeis queriedwithequalsnull shouldnotreturn hz vigorous jennings invocationmonitorthread retrying invocation invocation op com hazelcast map impl query queryoperation servicename hz impl mapservice identityhash partitionid replicaindex callid invocationtime waittimeout calltimeout tenantcontrol com hazelcast spi impl tenantcontrol nooptenantcontrol name trycount trypausemillis invokecount calltimeoutmillis firstinvocationtimems firstinvocationtime lastheartbeatmillis lastheartbeattime target null pendingresponse void backupsacksexpected backupsacksreceived connection null reason com hazelcast spi exception wrongtargetexception wrongtarget local member this expected target null partitionid replicaindex operation com hazelcast map impl query queryoperation service hz impl mapservice warn testnonterminalattributeis queriedwithequalsnull shouldnotreturn hz vigorous jennings invocationmonitorthread retrying invocation invocation op com hazelcast map impl query queryoperation servicename hz impl mapservice identityhash partitionid replicaindex callid invocationtime waittimeout calltimeout tenantcontrol com hazelcast spi impl tenantcontrol nooptenantcontrol name trycount trypausemillis invokecount calltimeoutmillis firstinvocationtimems firstinvocationtime lastheartbeatmillis lastheartbeattime target null pendingresponse void backupsacksexpected backupsacksreceived connection null reason com hazelcast spi exception wrongtargetexception wrongtarget local member this expected target null partitionid replicaindex operation com hazelcast map impl query queryoperation service hz impl mapservice it would be good to investigate the root cause of this
1
481,190
13,881,750,872
IssuesEvent
2020-10-18 02:23:50
apcountryman/avr-libcpp
https://api.github.com/repos/apcountryman/avr-libcpp
closed
Add util/delay
priority-normal status-in_revision type-feature
Add `util/delay` header (`system/util/delay`), and associated header/source implementation files (`include/util/delay.h`, and `source/util/delay.cc`).
1.0
Add util/delay - Add `util/delay` header (`system/util/delay`), and associated header/source implementation files (`include/util/delay.h`, and `source/util/delay.cc`).
non_defect
add util delay add util delay header system util delay and associated header source implementation files include util delay h and source util delay cc
0
21,573
3,520,005,485
IssuesEvent
2016-01-12 19:03:00
WilliamOckham/hunpos
https://api.github.com/repos/WilliamOckham/hunpos
closed
tagging erroneous in some cases
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? Tag a sample corpus What is the expected output? What do you see instead? I wrote in for example: got expected I PRP do VBP work NN <-- VB . SENT What version of the product are you using? On what operating system? 1.2.8, linux mint Please provide any additional information below. ``` Original issue reported on code.google.com by `krie...@gmx.de` on 19 Dec 2010 at 8:59
1.0
tagging erroneous in some cases - ``` What steps will reproduce the problem? Tag a sample corpus What is the expected output? What do you see instead? I wrote in for example: got expected I PRP do VBP work NN <-- VB . SENT What version of the product are you using? On what operating system? 1.2.8, linux mint Please provide any additional information below. ``` Original issue reported on code.google.com by `krie...@gmx.de` on 19 Dec 2010 at 8:59
defect
tagging erroneous in some cases what steps will reproduce the problem tag a sample corpus what is the expected output what do you see instead i wrote in for example got expected i prp do vbp work nn vb sent what version of the product are you using on what operating system linux mint please provide any additional information below original issue reported on code google com by krie gmx de on dec at
1
486,981
14,017,401,933
IssuesEvent
2020-10-29 15:36:47
GQCG/GQCP
https://api.github.com/repos/GQCG/GQCP
opened
Enable the evaluation of `GSQOperators` in a frozen-core spin-unresolved ONV basis
C++ complexity: intermediate priority: low theory
In recent refactors (#688), we had to temporarily disable some of the CI functionality. This issue tracks the re-enabling of the API to evaluate restricted operators in a seniority-zero ONV basis.
1.0
Enable the evaluation of `GSQOperators` in a frozen-core spin-unresolved ONV basis - In recent refactors (#688), we had to temporarily disable some of the CI functionality. This issue tracks the re-enabling of the API to evaluate restricted operators in a seniority-zero ONV basis.
non_defect
enable the evaluation of gsqoperators in a frozen core spin unresolved onv basis in recent refactors we had to temporarily disable some of the ci functionality this issue tracks the re enabling of the api to evaluate restricted operators in a seniority zero onv basis
0
28,303
5,239,168,291
IssuesEvent
2017-01-31 08:53:26
pexcn/tb-tun
https://api.github.com/repos/pexcn/tb-tun
closed
MTU and ICMP issue
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. ICMP package on IPv4 should be reflected to the IPv6 according to RFC2893. But TB-TUN has nothing to do with ICMP. 2. According to RFC3506, "If the IPv6 MTU size proves to be too large for some intermediate IPv4 subnet, IPv4 fragmentation will ensue....The IPv4 'do not fragment' bit *SHOULD NOT* be set in the encapsulating IPv4 header." But TB-TUN disobey these 2 rules. It causes IPv6 packages missing data and did not report any error by ICMP when IPv6 MTU is to large. What is the expected output? What do you see instead? Please use labels and text to provide additional information. ``` Original issue reported on code.google.com by `lostrive...@gmail.com` on 25 Apr 2010 at 5:50
1.0
MTU and ICMP issue - ``` What steps will reproduce the problem? 1. ICMP package on IPv4 should be reflected to the IPv6 according to RFC2893. But TB-TUN has nothing to do with ICMP. 2. According to RFC3506, "If the IPv6 MTU size proves to be too large for some intermediate IPv4 subnet, IPv4 fragmentation will ensue....The IPv4 'do not fragment' bit *SHOULD NOT* be set in the encapsulating IPv4 header." But TB-TUN disobey these 2 rules. It causes IPv6 packages missing data and did not report any error by ICMP when IPv6 MTU is to large. What is the expected output? What do you see instead? Please use labels and text to provide additional information. ``` Original issue reported on code.google.com by `lostrive...@gmail.com` on 25 Apr 2010 at 5:50
defect
mtu and icmp issue what steps will reproduce the problem icmp package on should be reflected to the according to but tb tun has nothing to do with icmp according to if the mtu size proves to be too large for some intermediate subnet fragmentation will ensue the do not fragment bit should not be set in the encapsulating header but tb tun disobey these rules it causes packages missing data and did not report any error by icmp when mtu is to large what is the expected output what do you see instead please use labels and text to provide additional information original issue reported on code google com by lostrive gmail com on apr at
1
3,866
2,610,083,167
IssuesEvent
2015-02-26 18:25:25
chrsmith/dsdsdaadf
https://api.github.com/repos/chrsmith/dsdsdaadf
opened
祛除青春痘深圳
auto-migrated Priority-Medium Type-Defect
``` 祛除青春痘深圳【深圳韩方科颜全国热线400-869-1818,24小时QQ4 008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘方�� �—韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方科� ��专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健康 祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业治�� �粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘痘� �� ``` ----- Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 6:47
1.0
祛除青春痘深圳 - ``` 祛除青春痘深圳【深圳韩方科颜全国热线400-869-1818,24小时QQ4 008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘方�� �—韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方科� ��专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健康 祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业治�� �粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘痘� �� ``` ----- Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 6:47
defect
祛除青春痘深圳 祛除青春痘深圳【 , 】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘方�� �—韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方科� ��专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健康 祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业治�� �粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘痘� �� original issue reported on code google com by szft com on may at
1
59,514
24,802,375,927
IssuesEvent
2022-10-24 23:18:13
Azure/azure-sdk-for-python
https://api.github.com/repos/Azure/azure-sdk-for-python
closed
ApplicationInsights does not work out of the box when using ClientSecretCredentials
feature-request Monitor Service Attention Client Monitor - ApplicationInsights
- **Package Name**: azure-applicationinsights - **Package Version**: 0.1.0 - **Operating System**: Linux - **Python Version**: 3.8.7 **Describe the bug** The applicationInsights client requires: - the identity credential must have a signed_session (that is missing from the azure.identity.ClientSecretCredential) so a [wrapper](https://gist.github.com/lmazuel/cc683d82ea1d7b40208de7c9fc8de59d) has to be used to make it work. - AppId and AppKey are needed to fetch queries succesfully: To pass the app key, the header needs to be modified in the applicationInsightDataClient ``` python credential = ClientSecretCredential( client_id=os.environ.get(env.ENV_APP_CLIENT_ID), client_secret=os.environ.get(env.ENV_APP_SECRET), tenant_id=os.environ.get(env.ENV_APP_TENANT), ) cred_adapter = CredentialWrapper(credential) client = ApplicationInsightsDataClient(cred_adapter) client.config.headers.setdefault("x-api-key", os.environ.get(env.ENV_APPLICATION_INSIGHTS_KEY)) ``` There should be a special Credential class for using ApplicationInsights without using wrappers and adding custom headers or ApplicationInsights should support authentication with ClientSecretCredentials only without requiring the Application Insights AppKey.
1.0
ApplicationInsights does not work out of the box when using ClientSecretCredentials - - **Package Name**: azure-applicationinsights - **Package Version**: 0.1.0 - **Operating System**: Linux - **Python Version**: 3.8.7 **Describe the bug** The applicationInsights client requires: - the identity credential must have a signed_session (that is missing from the azure.identity.ClientSecretCredential) so a [wrapper](https://gist.github.com/lmazuel/cc683d82ea1d7b40208de7c9fc8de59d) has to be used to make it work. - AppId and AppKey are needed to fetch queries succesfully: To pass the app key, the header needs to be modified in the applicationInsightDataClient ``` python credential = ClientSecretCredential( client_id=os.environ.get(env.ENV_APP_CLIENT_ID), client_secret=os.environ.get(env.ENV_APP_SECRET), tenant_id=os.environ.get(env.ENV_APP_TENANT), ) cred_adapter = CredentialWrapper(credential) client = ApplicationInsightsDataClient(cred_adapter) client.config.headers.setdefault("x-api-key", os.environ.get(env.ENV_APPLICATION_INSIGHTS_KEY)) ``` There should be a special Credential class for using ApplicationInsights without using wrappers and adding custom headers or ApplicationInsights should support authentication with ClientSecretCredentials only without requiring the Application Insights AppKey.
non_defect
applicationinsights does not work out of the box when using clientsecretcredentials package name azure applicationinsights package version operating system linux python version describe the bug the applicationinsights client requires the identity credential must have a signed session that is missing from the azure identity clientsecretcredential so a has to be used to make it work appid and appkey are needed to fetch queries succesfully to pass the app key the header needs to be modified in the applicationinsightdataclient python credential clientsecretcredential client id os environ get env env app client id client secret os environ get env env app secret tenant id os environ get env env app tenant cred adapter credentialwrapper credential client applicationinsightsdataclient cred adapter client config headers setdefault x api key os environ get env env application insights key there should be a special credential class for using applicationinsights without using wrappers and adding custom headers or applicationinsights should support authentication with clientsecretcredentials only without requiring the application insights appkey
0
16,685
2,931,469,148
IssuesEvent
2015-06-29 12:44:57
MarcusWolschon/osmeditor4android
https://api.github.com/repos/MarcusWolschon/osmeditor4android
closed
0.9.6 startup crash related to lock icon not being found
Defect FIXED!
A small number of crash dumps have been submitted that indicate an issue immediately after upgrade. Issue is due to deprecated modes being turned on, which in itself is deprecated.
1.0
0.9.6 startup crash related to lock icon not being found - A small number of crash dumps have been submitted that indicate an issue immediately after upgrade. Issue is due to deprecated modes being turned on, which in itself is deprecated.
defect
startup crash related to lock icon not being found a small number of crash dumps have been submitted that indicate an issue immediately after upgrade issue is due to deprecated modes being turned on which in itself is deprecated
1
159,936
6,064,642,121
IssuesEvent
2017-06-14 14:42:09
openvstorage/openvstorage-health-check
https://api.github.com/repos/openvstorage/openvstorage-health-check
closed
healthcheck never ends when checks are waiting on memcache
priority_minor type_bug
Best solution is to use a timeout frame when the check is waiting on memcache.
1.0
healthcheck never ends when checks are waiting on memcache - Best solution is to use a timeout frame when the check is waiting on memcache.
non_defect
healthcheck never ends when checks are waiting on memcache best solution is to use a timeout frame when the check is waiting on memcache
0
36,656
8,049,397,844
IssuesEvent
2018-08-01 09:58:05
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
TCP discovery with public-address specified doesn't handle correctly member shutdowns
Estimation: M Team: Core Type: Defect
I see "Connection reset by peer exceptions" in the (active) member log after one of Hazelcast members shuts down in my Docker environment. ### Environment Docker host runs 2 containers with Hazelcast members. Both have configured TCP discovery and hosts public address. The Hazelcast port is exposed to the Docker host. #### hazelcast.xml ```xml <port auto-increment="false" port-count="100">5701</port> <public-address>${hazelcast.public.ip}:${hazelcast.port}</public-address> <join> <multicast enabled="false"> </multicast> <tcp-ip enabled="true"> <member>${hazelcast.public.ip}:5555</member> <member>${hazelcast.public.ip}:5556</member> </tcp-ip> ... </join> ``` #### Start Docker containers ```bash docker run -it --rm -p 5555:5701 hazelcast-tcp docker run -it --rm -p 5556:5701 hazelcast-tcp ``` #### Start Hazelcast in the containers Provide system properties used in `hazelcast.xml` (public Docker host IP address and the port - 5555 and 5556): ``` System.setProperty("hazelcast.public.ip", "192.168.1.105"); // replace the IP value with your own IP System.setProperty("hazelcast.port", "5555"); // for the second container use 5556 hz = Hazelcast.newHazelcastInstance(); ``` #### Shut down one of the containers ```java hz.shutdown(); ``` #### Check the log in the running container The remaining cluster member try to connect to the stopped one and log contains bunch of entries similar to: ``` Oct 31, 2017 12:14:55 PM com.hazelcast.nio.tcp.TcpIpConnector INFO: [192.168.1.105]:5555 [dev] [3.9] Connecting to /192.168.1.105:5556, timeout: 0, bind-any: true Oct 31, 2017 12:14:55 PM com.hazelcast.nio.tcp.TcpIpConnectionManager INFO: [192.168.1.105]:5555 [dev] [3.9] Established socket connection between /172.17.0.3:56203 and /192.168.1.105:5556 Oct 31, 2017 12:14:55 PM com.hazelcast.nio.tcp.TcpIpConnection INFO: [192.168.1.105]:5555 [dev] [3.9] Connection[id=24, /172.17.0.3:56203->/192.168.1.105:5556, endpoint=[192.168.1.105]:5556, alive=false, type=NONE] closed. Reason: Connection closed by the other side Oct 31, 2017 12:14:55 PM com.hazelcast.nio.tcp.TcpIpConnector INFO: [192.168.1.105]:5555 [dev] [3.9] Connecting to /192.168.1.105:5556, timeout: 0, bind-any: true Oct 31, 2017 12:14:55 PM com.hazelcast.nio.tcp.TcpIpConnectionManager INFO: [192.168.1.105]:5555 [dev] [3.9] Established socket connection between /172.17.0.3:53561 and /192.168.1.105:5556 Oct 31, 2017 12:14:55 PM com.hazelcast.nio.tcp.TcpIpConnection WARNING: [192.168.1.105]:5555 [dev] [3.9] Connection[id=25, /172.17.0.3:53561->/192.168.1.105:5556, endpoint=[192.168.1.105]:5556, alive=false, type=NONE] closed. Reason: Exception in Connection[id=25, /172.17.0.3:53561->/192.168.1.105:5556, endpoint=[192.168.1.105]:5556, alive=true, type=NONE], thread=hz._hzInstance_1_dev.IO.thread-in-0 java.io.IOException: Connection reset by peer at sun.nio.ch.FileDispatcherImpl.read0(Native Method) at sun.nio.ch.SocketDispatcher.read(SocketDispatcher.java:39) at sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:223) at sun.nio.ch.IOUtil.read(IOUtil.java:197) at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:380) at com.hazelcast.internal.networking.AbstractChannel.read(AbstractChannel.java:94) at com.hazelcast.nio.tcp.MemberChannelInitializer.inboundProtocol(MemberChannelInitializer.java:101) at com.hazelcast.nio.tcp.MemberChannelInitializer.initInbound(MemberChannelInitializer.java:73) at com.hazelcast.internal.networking.nio.NioChannelReader.init(NioChannelReader.java:143) at com.hazelcast.internal.networking.nio.NioChannelReader.handle(NioChannelReader.java:123) at com.hazelcast.internal.networking.nio.NioThread.handleSelectionKey(NioThread.java:401) at com.hazelcast.internal.networking.nio.NioThread.handleSelectionKeys(NioThread.java:386) at com.hazelcast.internal.networking.nio.NioThread.selectLoop(NioThread.java:293) at com.hazelcast.internal.networking.nio.NioThread.run(NioThread.java:248) ```
1.0
TCP discovery with public-address specified doesn't handle correctly member shutdowns - I see "Connection reset by peer exceptions" in the (active) member log after one of Hazelcast members shuts down in my Docker environment. ### Environment Docker host runs 2 containers with Hazelcast members. Both have configured TCP discovery and hosts public address. The Hazelcast port is exposed to the Docker host. #### hazelcast.xml ```xml <port auto-increment="false" port-count="100">5701</port> <public-address>${hazelcast.public.ip}:${hazelcast.port}</public-address> <join> <multicast enabled="false"> </multicast> <tcp-ip enabled="true"> <member>${hazelcast.public.ip}:5555</member> <member>${hazelcast.public.ip}:5556</member> </tcp-ip> ... </join> ``` #### Start Docker containers ```bash docker run -it --rm -p 5555:5701 hazelcast-tcp docker run -it --rm -p 5556:5701 hazelcast-tcp ``` #### Start Hazelcast in the containers Provide system properties used in `hazelcast.xml` (public Docker host IP address and the port - 5555 and 5556): ``` System.setProperty("hazelcast.public.ip", "192.168.1.105"); // replace the IP value with your own IP System.setProperty("hazelcast.port", "5555"); // for the second container use 5556 hz = Hazelcast.newHazelcastInstance(); ``` #### Shut down one of the containers ```java hz.shutdown(); ``` #### Check the log in the running container The remaining cluster member try to connect to the stopped one and log contains bunch of entries similar to: ``` Oct 31, 2017 12:14:55 PM com.hazelcast.nio.tcp.TcpIpConnector INFO: [192.168.1.105]:5555 [dev] [3.9] Connecting to /192.168.1.105:5556, timeout: 0, bind-any: true Oct 31, 2017 12:14:55 PM com.hazelcast.nio.tcp.TcpIpConnectionManager INFO: [192.168.1.105]:5555 [dev] [3.9] Established socket connection between /172.17.0.3:56203 and /192.168.1.105:5556 Oct 31, 2017 12:14:55 PM com.hazelcast.nio.tcp.TcpIpConnection INFO: [192.168.1.105]:5555 [dev] [3.9] Connection[id=24, /172.17.0.3:56203->/192.168.1.105:5556, endpoint=[192.168.1.105]:5556, alive=false, type=NONE] closed. Reason: Connection closed by the other side Oct 31, 2017 12:14:55 PM com.hazelcast.nio.tcp.TcpIpConnector INFO: [192.168.1.105]:5555 [dev] [3.9] Connecting to /192.168.1.105:5556, timeout: 0, bind-any: true Oct 31, 2017 12:14:55 PM com.hazelcast.nio.tcp.TcpIpConnectionManager INFO: [192.168.1.105]:5555 [dev] [3.9] Established socket connection between /172.17.0.3:53561 and /192.168.1.105:5556 Oct 31, 2017 12:14:55 PM com.hazelcast.nio.tcp.TcpIpConnection WARNING: [192.168.1.105]:5555 [dev] [3.9] Connection[id=25, /172.17.0.3:53561->/192.168.1.105:5556, endpoint=[192.168.1.105]:5556, alive=false, type=NONE] closed. Reason: Exception in Connection[id=25, /172.17.0.3:53561->/192.168.1.105:5556, endpoint=[192.168.1.105]:5556, alive=true, type=NONE], thread=hz._hzInstance_1_dev.IO.thread-in-0 java.io.IOException: Connection reset by peer at sun.nio.ch.FileDispatcherImpl.read0(Native Method) at sun.nio.ch.SocketDispatcher.read(SocketDispatcher.java:39) at sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:223) at sun.nio.ch.IOUtil.read(IOUtil.java:197) at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:380) at com.hazelcast.internal.networking.AbstractChannel.read(AbstractChannel.java:94) at com.hazelcast.nio.tcp.MemberChannelInitializer.inboundProtocol(MemberChannelInitializer.java:101) at com.hazelcast.nio.tcp.MemberChannelInitializer.initInbound(MemberChannelInitializer.java:73) at com.hazelcast.internal.networking.nio.NioChannelReader.init(NioChannelReader.java:143) at com.hazelcast.internal.networking.nio.NioChannelReader.handle(NioChannelReader.java:123) at com.hazelcast.internal.networking.nio.NioThread.handleSelectionKey(NioThread.java:401) at com.hazelcast.internal.networking.nio.NioThread.handleSelectionKeys(NioThread.java:386) at com.hazelcast.internal.networking.nio.NioThread.selectLoop(NioThread.java:293) at com.hazelcast.internal.networking.nio.NioThread.run(NioThread.java:248) ```
defect
tcp discovery with public address specified doesn t handle correctly member shutdowns i see connection reset by peer exceptions in the active member log after one of hazelcast members shuts down in my docker environment environment docker host runs containers with hazelcast members both have configured tcp discovery and hosts public address the hazelcast port is exposed to the docker host hazelcast xml xml hazelcast public ip hazelcast port hazelcast public ip hazelcast public ip start docker containers bash docker run it rm p hazelcast tcp docker run it rm p hazelcast tcp start hazelcast in the containers provide system properties used in hazelcast xml public docker host ip address and the port and system setproperty hazelcast public ip replace the ip value with your own ip system setproperty hazelcast port for the second container use hz hazelcast newhazelcastinstance shut down one of the containers java hz shutdown check the log in the running container the remaining cluster member try to connect to the stopped one and log contains bunch of entries similar to oct pm com hazelcast nio tcp tcpipconnector info connecting to timeout bind any true oct pm com hazelcast nio tcp tcpipconnectionmanager info established socket connection between and oct pm com hazelcast nio tcp tcpipconnection info connection alive false type none closed reason connection closed by the other side oct pm com hazelcast nio tcp tcpipconnector info connecting to timeout bind any true oct pm com hazelcast nio tcp tcpipconnectionmanager info established socket connection between and oct pm com hazelcast nio tcp tcpipconnection warning connection alive false type none closed reason exception in connection alive true type none thread hz hzinstance dev io thread in java io ioexception connection reset by peer at sun nio ch filedispatcherimpl native method at sun nio ch socketdispatcher read socketdispatcher java at sun nio ch ioutil readintonativebuffer ioutil java at sun nio ch ioutil read ioutil java at sun nio ch socketchannelimpl read socketchannelimpl java at com hazelcast internal networking abstractchannel read abstractchannel java at com hazelcast nio tcp memberchannelinitializer inboundprotocol memberchannelinitializer java at com hazelcast nio tcp memberchannelinitializer initinbound memberchannelinitializer java at com hazelcast internal networking nio niochannelreader init niochannelreader java at com hazelcast internal networking nio niochannelreader handle niochannelreader java at com hazelcast internal networking nio niothread handleselectionkey niothread java at com hazelcast internal networking nio niothread handleselectionkeys niothread java at com hazelcast internal networking nio niothread selectloop niothread java at com hazelcast internal networking nio niothread run niothread java
1
484,282
13,937,402,661
IssuesEvent
2020-10-22 14:06:34
ViRGiL175/java-diner-automation
https://api.github.com/repos/ViRGiL175/java-diner-automation
opened
Босс видит фидбек клиентов через интерфейс
complexity:_3 priority:_1 type:_story
Босс хочет наблюдать за фидбеком клиентов дайнера через удобный и понятный интерфейс приложения, чтобы знать, насколько клиенты довольны обслуживанием ## Описание > Описание задачи, схемы, картинки и т.д. ## Критерии выполненности - [ ] Босс видит фидбек клиентов - [ ] Данные регулярно обновляются ## Связанное > Всяческие подробности, ссылки, документация и т.д.
1.0
Босс видит фидбек клиентов через интерфейс - Босс хочет наблюдать за фидбеком клиентов дайнера через удобный и понятный интерфейс приложения, чтобы знать, насколько клиенты довольны обслуживанием ## Описание > Описание задачи, схемы, картинки и т.д. ## Критерии выполненности - [ ] Босс видит фидбек клиентов - [ ] Данные регулярно обновляются ## Связанное > Всяческие подробности, ссылки, документация и т.д.
non_defect
босс видит фидбек клиентов через интерфейс босс хочет наблюдать за фидбеком клиентов дайнера через удобный и понятный интерфейс приложения чтобы знать насколько клиенты довольны обслуживанием описание описание задачи схемы картинки и т д критерии выполненности босс видит фидбек клиентов данные регулярно обновляются связанное всяческие подробности ссылки документация и т д
0
66,248
20,103,291,856
IssuesEvent
2022-02-07 07:54:56
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
Compact GenericRecords' getFieldKind should throw [HZ-879]
Type: Defect Team: Client Source: Internal Estimation: S Module: Serialization to-jira
When field name does not exist in generic record according to this code: https://github.com/hazelcast/hazelcast/blob/0ab9eb067f04cbbb62d8998ee21c99cb05239761/hazelcast/src/main/java/com/hazelcast/nio/serialization/GenericRecord.java#L119-L125 getFieldKind should throw and should not return null. (NonNull annotation) However, as can be seen from https://github.com/hazelcast/hazelcast/blob/77ef2bfdfb19005b560cefb2a104994019983d2a/hazelcast/src/main/java/com/hazelcast/internal/serialization/impl/compact/DeserializedGenericRecord.java#L107-L111 and https://github.com/hazelcast/hazelcast/blob/4544e4a6fa5060fa9ae2984a20ed3956099cc583/hazelcast/src/main/java/com/hazelcast/internal/serialization/impl/compact/CompactInternalGenericRecord.java#L165-L169 they don't throw.
1.0
Compact GenericRecords' getFieldKind should throw [HZ-879] - When field name does not exist in generic record according to this code: https://github.com/hazelcast/hazelcast/blob/0ab9eb067f04cbbb62d8998ee21c99cb05239761/hazelcast/src/main/java/com/hazelcast/nio/serialization/GenericRecord.java#L119-L125 getFieldKind should throw and should not return null. (NonNull annotation) However, as can be seen from https://github.com/hazelcast/hazelcast/blob/77ef2bfdfb19005b560cefb2a104994019983d2a/hazelcast/src/main/java/com/hazelcast/internal/serialization/impl/compact/DeserializedGenericRecord.java#L107-L111 and https://github.com/hazelcast/hazelcast/blob/4544e4a6fa5060fa9ae2984a20ed3956099cc583/hazelcast/src/main/java/com/hazelcast/internal/serialization/impl/compact/CompactInternalGenericRecord.java#L165-L169 they don't throw.
defect
compact genericrecords getfieldkind should throw when field name does not exist in generic record according to this code getfieldkind should throw and should not return null nonnull annotation however as can be seen from and they don t throw
1
36,687
12,418,576,432
IssuesEvent
2020-05-23 01:02:00
wrbejar/Autofac
https://api.github.com/repos/wrbejar/Autofac
opened
CVE-2020-7656 (Medium) detected in jquery-1.4.1.min.js, jquery-1.4.1.js
security vulnerability
## CVE-2020-7656 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-1.4.1.min.js</b>, <b>jquery-1.4.1.js</b></p></summary> <p> <details><summary><b>jquery-1.4.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.1/jquery.min.js</a></p> <p>Path to vulnerable library: /Autofac/src/MultitenantExample.MvcApplication/Scripts/jquery-1.4.1.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.4.1.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-1.4.1.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.1/jquery.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.1/jquery.js</a></p> <p>Path to vulnerable library: /Autofac/src/MultitenantExample.MvcApplication/Scripts/jquery-1.4.1.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.4.1.js** (Vulnerable Library) </details> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jquery prior to 1.9.0 allows Cross-site Scripting attacks via the load method. The load method fails to recognize and remove "<script>" HTML tags that contain a whitespace character, i.e: "</script >", which results in the enclosed script logic to be executed. <p>Publish Date: 2020-05-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7656>CVE-2020-7656</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7656">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7656</a></p> <p>Release Date: 2020-05-19</p> <p>Fix Resolution: 1.9.0b1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.4.1","isTransitiveDependency":false,"dependencyTree":"jquery:1.4.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.9.0b1"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.4.1","isTransitiveDependency":false,"dependencyTree":"jquery:1.4.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.9.0b1"}],"vulnerabilityIdentifier":"CVE-2020-7656","vulnerabilityDetails":"jquery prior to 1.9.0 allows Cross-site Scripting attacks via the load method. The load method fails to recognize and remove \"\u003cscript\u003e\" HTML tags that contain a whitespace character, i.e: \"\u003c/script \u003e\", which results in the enclosed script logic to be executed.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7656","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-7656 (Medium) detected in jquery-1.4.1.min.js, jquery-1.4.1.js - ## CVE-2020-7656 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-1.4.1.min.js</b>, <b>jquery-1.4.1.js</b></p></summary> <p> <details><summary><b>jquery-1.4.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.1/jquery.min.js</a></p> <p>Path to vulnerable library: /Autofac/src/MultitenantExample.MvcApplication/Scripts/jquery-1.4.1.min.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.4.1.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-1.4.1.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.1/jquery.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.1/jquery.js</a></p> <p>Path to vulnerable library: /Autofac/src/MultitenantExample.MvcApplication/Scripts/jquery-1.4.1.js</p> <p> Dependency Hierarchy: - :x: **jquery-1.4.1.js** (Vulnerable Library) </details> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jquery prior to 1.9.0 allows Cross-site Scripting attacks via the load method. The load method fails to recognize and remove "<script>" HTML tags that contain a whitespace character, i.e: "</script >", which results in the enclosed script logic to be executed. <p>Publish Date: 2020-05-19 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7656>CVE-2020-7656</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7656">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7656</a></p> <p>Release Date: 2020-05-19</p> <p>Fix Resolution: 1.9.0b1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.4.1","isTransitiveDependency":false,"dependencyTree":"jquery:1.4.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.9.0b1"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.4.1","isTransitiveDependency":false,"dependencyTree":"jquery:1.4.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.9.0b1"}],"vulnerabilityIdentifier":"CVE-2020-7656","vulnerabilityDetails":"jquery prior to 1.9.0 allows Cross-site Scripting attacks via the load method. The load method fails to recognize and remove \"\u003cscript\u003e\" HTML tags that contain a whitespace character, i.e: \"\u003c/script \u003e\", which results in the enclosed script logic to be executed.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7656","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_defect
cve medium detected in jquery min js jquery js cve medium severity vulnerability vulnerable libraries jquery min js jquery js jquery min js javascript library for dom operations library home page a href path to vulnerable library autofac src multitenantexample mvcapplication scripts jquery min js dependency hierarchy x jquery min js vulnerable library jquery js javascript library for dom operations library home page a href path to vulnerable library autofac src multitenantexample mvcapplication scripts jquery js dependency hierarchy x jquery js vulnerable library vulnerability details jquery prior to allows cross site scripting attacks via the load method the load method fails to recognize and remove html tags that contain a whitespace character i e which results in the enclosed script logic to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails jquery prior to allows cross site scripting attacks via the load method the load method fails to recognize and remove html tags that contain a whitespace character i e script which results in the enclosed script logic to be executed vulnerabilityurl
0
44,399
12,132,757,919
IssuesEvent
2020-04-23 07:52:33
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
opened
hazelcast-client:4.0 is missing
Type: Defect
Looks like the hazelcast-client 4.X was not released. For the Hazelcast server and Hazelcast-hibernate5 artefacts we can migrate to Hazelcast 4 but there is not client artefact at Maven central (or JCenter). [Your documentation](https://hazelcast.org/imdg/clients-languages/java/) already mentions version 4.0.1 of the hazelcast-client but I can not find it anywhere.
1.0
hazelcast-client:4.0 is missing - Looks like the hazelcast-client 4.X was not released. For the Hazelcast server and Hazelcast-hibernate5 artefacts we can migrate to Hazelcast 4 but there is not client artefact at Maven central (or JCenter). [Your documentation](https://hazelcast.org/imdg/clients-languages/java/) already mentions version 4.0.1 of the hazelcast-client but I can not find it anywhere.
defect
hazelcast client is missing looks like the hazelcast client x was not released for the hazelcast server and hazelcast artefacts we can migrate to hazelcast but there is not client artefact at maven central or jcenter already mentions version of the hazelcast client but i can not find it anywhere
1
44,915
12,440,273,664
IssuesEvent
2020-05-26 11:41:21
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
closed
Native tooltip [title] does not change for FileUpload component in advanced mode
defect
By default native HTML `input type="file"` will have a tolltip specific to browser. On Firefox its: _No file selected._ On chrome its: _No file chosen_ After a file has been selected by user, the tolltip will change to selected file name. On Primefaces FileUpload with Basic mode this works just as expected. But when used as Advanced mode the tooltip will never change, it will always be same. Apart from this issue, it would be good to have tooltip attribute for Fileupload component.
1.0
Native tooltip [title] does not change for FileUpload component in advanced mode - By default native HTML `input type="file"` will have a tolltip specific to browser. On Firefox its: _No file selected._ On chrome its: _No file chosen_ After a file has been selected by user, the tolltip will change to selected file name. On Primefaces FileUpload with Basic mode this works just as expected. But when used as Advanced mode the tooltip will never change, it will always be same. Apart from this issue, it would be good to have tooltip attribute for Fileupload component.
defect
native tooltip does not change for fileupload component in advanced mode by default native html input type file will have a tolltip specific to browser on firefox its no file selected on chrome its no file chosen after a file has been selected by user the tolltip will change to selected file name on primefaces fileupload with basic mode this works just as expected but when used as advanced mode the tooltip will never change it will always be same apart from this issue it would be good to have tooltip attribute for fileupload component
1
139,262
20,820,627,113
IssuesEvent
2022-03-18 15:01:47
cuzekn/My-Best-Book
https://api.github.com/repos/cuzekn/My-Best-Book
opened
ユーザー認証(ユーザー作成画面とログイン画面)を実装
design
[やること] - Reactアプリの環境構築 - ルーティングの作成 - ユーザー作成画面の作成(/signup) URL例) localhost/signup - ログイン画面の作成(/login) URL例) localhost/login
1.0
ユーザー認証(ユーザー作成画面とログイン画面)を実装 - [やること] - Reactアプリの環境構築 - ルーティングの作成 - ユーザー作成画面の作成(/signup) URL例) localhost/signup - ログイン画面の作成(/login) URL例) localhost/login
non_defect
ユーザー認証 ユーザー作成画面とログイン画面 を実装 reactアプリの環境構築 ルーティングの作成 ユーザー作成画面の作成 signup url例 localhost signup ログイン画面の作成 login url例 localhost login
0
30,606
6,192,417,563
IssuesEvent
2017-07-05 01:38:53
cakephp/cakephp
https://api.github.com/repos/cakephp/cakephp
closed
translate behaviour doesn't seem to patch translations on edit
behaviors Defect i18n On hold
This is a (multiple allowed): * [x] bug * [ ] enhancement * [ ] feature-discussion (RFC) * CakePHP Version: 3.4.7 * Platform and Target: ubuntu 16.04 ### What you did I created form fields using as follows: `_translations.en_GB.name` etc.. according to the docs here: https://book.cakephp.org/3.0/en/orm/behaviors/translate.html#saving-multiple-translations ### What happened when I use ```php $entity = $this->ModelTable->find('translations')->where(['id'=>'1]->first(); $this->ModelTable->patchEntity($entity, $this->request->getData(), ['translations'=>true]); ``` translated fields are not patched even though the data from the form seems to be correct. ### What you expected to happen I would expect translated fields to be patched as well
1.0
translate behaviour doesn't seem to patch translations on edit - This is a (multiple allowed): * [x] bug * [ ] enhancement * [ ] feature-discussion (RFC) * CakePHP Version: 3.4.7 * Platform and Target: ubuntu 16.04 ### What you did I created form fields using as follows: `_translations.en_GB.name` etc.. according to the docs here: https://book.cakephp.org/3.0/en/orm/behaviors/translate.html#saving-multiple-translations ### What happened when I use ```php $entity = $this->ModelTable->find('translations')->where(['id'=>'1]->first(); $this->ModelTable->patchEntity($entity, $this->request->getData(), ['translations'=>true]); ``` translated fields are not patched even though the data from the form seems to be correct. ### What you expected to happen I would expect translated fields to be patched as well
defect
translate behaviour doesn t seem to patch translations on edit this is a multiple allowed bug enhancement feature discussion rfc cakephp version platform and target ubuntu what you did i created form fields using as follows translations en gb name etc according to the docs here what happened when i use php entity this modeltable find translations where first this modeltable patchentity entity this request getdata translated fields are not patched even though the data from the form seems to be correct what you expected to happen i would expect translated fields to be patched as well
1
71,147
23,469,347,357
IssuesEvent
2022-08-16 20:04:02
department-of-veterans-affairs/va.gov-cms
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
opened
Duplicate calls to action on Vet Centers Dashboard cards
Needs refining ⭐️ Sitewide CMS 508/Accessibility 508-defect-4
## Description On the Vet Center Dashboard in the CMS, there are multiple cards that each have a call to action for "View" and "Edit", causing duplicate links. Ideally each link should have distinct, unique text to indicate to the user where the link would direct them - "Edit Locations Page" or "View Main Vet Center Page" ## Screenshot ![image](https://user-images.githubusercontent.com/106678594/184974308-db9c7d4c-6611-4f1f-aad3-c35c096e193f.png) ## Accessibility Standard WCAG version 2.0 AAA, [Criterion 2.4.9](https://www.w3.org/WAI/WCAG21/Understanding/link-purpose-link-only.html) ## Acceptance Criteria - [ ] UX review to determine content updates for links - [ ] Technical review - [ ] Change management consulted - [ ] Implementation ticket created ### CMS Team Please check the team(s) that will do this work. - [ ] `Program` - [ ] `Platform CMS Team` - [ ] `Sitewide Crew` - [ ] `⭐️ Sitewide CMS` - [ ] `⭐️ Public Websites` - [ ] `⭐️ Facilities` - [ ] `⭐️ User support`
1.0
Duplicate calls to action on Vet Centers Dashboard cards - ## Description On the Vet Center Dashboard in the CMS, there are multiple cards that each have a call to action for "View" and "Edit", causing duplicate links. Ideally each link should have distinct, unique text to indicate to the user where the link would direct them - "Edit Locations Page" or "View Main Vet Center Page" ## Screenshot ![image](https://user-images.githubusercontent.com/106678594/184974308-db9c7d4c-6611-4f1f-aad3-c35c096e193f.png) ## Accessibility Standard WCAG version 2.0 AAA, [Criterion 2.4.9](https://www.w3.org/WAI/WCAG21/Understanding/link-purpose-link-only.html) ## Acceptance Criteria - [ ] UX review to determine content updates for links - [ ] Technical review - [ ] Change management consulted - [ ] Implementation ticket created ### CMS Team Please check the team(s) that will do this work. - [ ] `Program` - [ ] `Platform CMS Team` - [ ] `Sitewide Crew` - [ ] `⭐️ Sitewide CMS` - [ ] `⭐️ Public Websites` - [ ] `⭐️ Facilities` - [ ] `⭐️ User support`
defect
duplicate calls to action on vet centers dashboard cards description on the vet center dashboard in the cms there are multiple cards that each have a call to action for view and edit causing duplicate links ideally each link should have distinct unique text to indicate to the user where the link would direct them edit locations page or view main vet center page screenshot accessibility standard wcag version aaa acceptance criteria ux review to determine content updates for links technical review change management consulted implementation ticket created cms team please check the team s that will do this work program platform cms team sitewide crew ⭐️ sitewide cms ⭐️ public websites ⭐️ facilities ⭐️ user support
1
331,009
10,058,746,177
IssuesEvent
2019-07-22 14:31:18
trailofbits/manticore
https://api.github.com/repos/trailofbits/manticore
opened
Assertion failure: assert value.size == self.size when gas is ignored
ethereum help wanted priority-high
### OS / Environment Ubuntu 18.04 ### Manticore version cc7811b10d9ce4f5c3444ebceff16b3ed2753f21 ### Python version Python 3.6 ### Summary of the problem Manticode crashes with an assertion failure when gas computation is ignored. ### Step to reproduce the behavior Use this solidity contract: ``` contract C { function hash(string x) public returns (bytes32) { return(keccak256(x)); } } ``` with manticore: ``` $ manticore bug.sol --txlimit 1 --evm.oog ignore ``` ### Expected behavior It should should not crash. It's unclear how to proceed with the symbolic exploration since manticore will waste resourced exploring the `keccak256` code. Maybe a warning saying that such path will not be explored? ### Actual behavior ``` 2019-07-22 10:24:54,158: [15617] m.main:INFO: Registered plugins: DetectReentrancyAdvanced, DetectSuicidal, DetectUnusedRetVal, DetectInvalid, DetectUninitializedMemory, DetectIntegerOverflow, DetectEnvInstruction, DetectExternalCallAndLeak, DetectDelegatecall, DetectReentrancySimple, DetectUninitializedStorage 2019-07-22 10:24:54,158: [15617] m.main:INFO: Beginning analysis 2019-07-22 10:24:54,163: [15617] m.e.manticore:INFO: Starting symbolic create contract 2019-07-22 10:24:54,710: [15617] m.e.manticore:INFO: Starting symbolic transaction: 0 2019-07-22 10:25:10,260: [15916] m.c.worker:ERROR: Exception in state 0: AssertionError() Traceback (most recent call last): File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/core/worker.py", line 121, in run current_state.execute() File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/ethereum/state.py", line 6, in execute return self._platform.execute() File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/platforms/evm.py", line 2684, in execute self.current_vm.execute() File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/platforms/evm.py", line 1210, in execute last_pc, last_gas, instruction, arguments, fee, allocated = self._checkpoint() File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/platforms/evm.py", line 1106, in _checkpoint fee = self._calculate_gas(*arguments) File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/platforms/evm.py", line 1079, in _calculate_gas return current.fee + implementation(*arguments) File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/platforms/evm.py", line 1533, in SHA3_gas return GSHA3WORD * (ceil32(size) // 32) + memfee File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/core/smtlib/expression.py", line 295, in __add__ return BitVecAdd(self, self.cast(other)) File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/core/smtlib/expression.py", line 284, in cast assert value.size == self.size AssertionError 2019-07-22 10:25:10,346: [15617] m.e.manticore:INFO: 0 alive states, 3 terminated states 2019-07-22 10:25:10,378: [16754] m.c.manticore:INFO: Generated testcase No. 0 - RETURN(1 txs) 2019-07-22 10:25:10,383: [16755] m.c.manticore:INFO: Generated testcase No. 1 - REVERT(2 txs) 2019-07-22 10:25:10,387: [16757] m.c.manticore:INFO: Generated testcase No. 2 - REVERT(2 txs) 2019-07-22 10:25:10,417: [16760] m.c.manticore:INFO: Generated testcase No. 3 - RETURN(1 txs) ```
1.0
Assertion failure: assert value.size == self.size when gas is ignored - ### OS / Environment Ubuntu 18.04 ### Manticore version cc7811b10d9ce4f5c3444ebceff16b3ed2753f21 ### Python version Python 3.6 ### Summary of the problem Manticode crashes with an assertion failure when gas computation is ignored. ### Step to reproduce the behavior Use this solidity contract: ``` contract C { function hash(string x) public returns (bytes32) { return(keccak256(x)); } } ``` with manticore: ``` $ manticore bug.sol --txlimit 1 --evm.oog ignore ``` ### Expected behavior It should should not crash. It's unclear how to proceed with the symbolic exploration since manticore will waste resourced exploring the `keccak256` code. Maybe a warning saying that such path will not be explored? ### Actual behavior ``` 2019-07-22 10:24:54,158: [15617] m.main:INFO: Registered plugins: DetectReentrancyAdvanced, DetectSuicidal, DetectUnusedRetVal, DetectInvalid, DetectUninitializedMemory, DetectIntegerOverflow, DetectEnvInstruction, DetectExternalCallAndLeak, DetectDelegatecall, DetectReentrancySimple, DetectUninitializedStorage 2019-07-22 10:24:54,158: [15617] m.main:INFO: Beginning analysis 2019-07-22 10:24:54,163: [15617] m.e.manticore:INFO: Starting symbolic create contract 2019-07-22 10:24:54,710: [15617] m.e.manticore:INFO: Starting symbolic transaction: 0 2019-07-22 10:25:10,260: [15916] m.c.worker:ERROR: Exception in state 0: AssertionError() Traceback (most recent call last): File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/core/worker.py", line 121, in run current_state.execute() File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/ethereum/state.py", line 6, in execute return self._platform.execute() File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/platforms/evm.py", line 2684, in execute self.current_vm.execute() File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/platforms/evm.py", line 1210, in execute last_pc, last_gas, instruction, arguments, fee, allocated = self._checkpoint() File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/platforms/evm.py", line 1106, in _checkpoint fee = self._calculate_gas(*arguments) File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/platforms/evm.py", line 1079, in _calculate_gas return current.fee + implementation(*arguments) File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/platforms/evm.py", line 1533, in SHA3_gas return GSHA3WORD * (ceil32(size) // 32) + memfee File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/core/smtlib/expression.py", line 295, in __add__ return BitVecAdd(self, self.cast(other)) File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/core/smtlib/expression.py", line 284, in cast assert value.size == self.size AssertionError 2019-07-22 10:25:10,346: [15617] m.e.manticore:INFO: 0 alive states, 3 terminated states 2019-07-22 10:25:10,378: [16754] m.c.manticore:INFO: Generated testcase No. 0 - RETURN(1 txs) 2019-07-22 10:25:10,383: [16755] m.c.manticore:INFO: Generated testcase No. 1 - REVERT(2 txs) 2019-07-22 10:25:10,387: [16757] m.c.manticore:INFO: Generated testcase No. 2 - REVERT(2 txs) 2019-07-22 10:25:10,417: [16760] m.c.manticore:INFO: Generated testcase No. 3 - RETURN(1 txs) ```
non_defect
assertion failure assert value size self size when gas is ignored os environment ubuntu manticore version python version python summary of the problem manticode crashes with an assertion failure when gas computation is ignored step to reproduce the behavior use this solidity contract contract c function hash string x public returns return x with manticore manticore bug sol txlimit evm oog ignore expected behavior it should should not crash it s unclear how to proceed with the symbolic exploration since manticore will waste resourced exploring the code maybe a warning saying that such path will not be explored actual behavior m main info registered plugins detectreentrancyadvanced detectsuicidal detectunusedretval detectinvalid detectuninitializedmemory detectintegeroverflow detectenvinstruction detectexternalcallandleak detectdelegatecall detectreentrancysimple detectuninitializedstorage m main info beginning analysis m e manticore info starting symbolic create contract m e manticore info starting symbolic transaction m c worker error exception in state assertionerror traceback most recent call last file home gustavo local lib site packages manticore egg manticore core worker py line in run current state execute file home gustavo local lib site packages manticore egg manticore ethereum state py line in execute return self platform execute file home gustavo local lib site packages manticore egg manticore platforms evm py line in execute self current vm execute file home gustavo local lib site packages manticore egg manticore platforms evm py line in execute last pc last gas instruction arguments fee allocated self checkpoint file home gustavo local lib site packages manticore egg manticore platforms evm py line in checkpoint fee self calculate gas arguments file home gustavo local lib site packages manticore egg manticore platforms evm py line in calculate gas return current fee implementation arguments file home gustavo local lib site packages manticore egg manticore platforms evm py line in gas return size memfee file home gustavo local lib site packages manticore egg manticore core smtlib expression py line in add return bitvecadd self self cast other file home gustavo local lib site packages manticore egg manticore core smtlib expression py line in cast assert value size self size assertionerror m e manticore info alive states terminated states m c manticore info generated testcase no return txs m c manticore info generated testcase no revert txs m c manticore info generated testcase no revert txs m c manticore info generated testcase no return txs
0
204,769
23,279,881,428
IssuesEvent
2022-08-05 10:57:00
Satheesh575555/Openssl_1_1_0_CVE-2021-23841
https://api.github.com/repos/Satheesh575555/Openssl_1_1_0_CVE-2021-23841
opened
CVE-2017-3738 (Medium) detected in opensslOpenSSL_1_1_0g
security vulnerability
## CVE-2017-3738 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>opensslOpenSSL_1_1_0g</b></p></summary> <p> <p>Akamai fork of openssl master.</p> <p>Library home page: <a href=https://github.com/akamai/openssl.git>https://github.com/akamai/openssl.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/Satheesh575555/Openssl_1_1_0_CVE-2021-23841/commit/4c0cda99bc44e45f32ea15e98aeb0a895ad1dd56">4c0cda99bc44e45f32ea15e98aeb0a895ad1dd56</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/crypto/bn/asm/rsaz-avx2.pl</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> There is an overflow bug in the AVX2 Montgomery multiplication procedure used in exponentiation with 1024-bit moduli. No EC algorithms are affected. Analysis suggests that attacks against RSA and DSA as a result of this defect would be very difficult to perform and are not believed likely. Attacks against DH1024 are considered just feasible, because most of the work necessary to deduce information about a private key may be performed offline. The amount of resources required for such an attack would be significant. However, for an attack on TLS to be meaningful, the server would have to share the DH1024 private key among multiple clients, which is no longer an option since CVE-2016-0701. This only affects processors that support the AVX2 but not ADX extensions like Intel Haswell (4th generation). Note: The impact from this issue is similar to CVE-2017-3736, CVE-2017-3732 and CVE-2015-3193. OpenSSL version 1.0.2-1.0.2m and 1.1.0-1.1.0g are affected. Fixed in OpenSSL 1.0.2n. Due to the low severity of this issue we are not issuing a new release of OpenSSL 1.1.0 at this time. The fix will be included in OpenSSL 1.1.0h when it becomes available. The fix is also available in commit e502cc86d in the OpenSSL git repository. <p>Publish Date: 2017-12-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-3738>CVE-2017-3738</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://gitlab.alpinelinux.org/alpine/aports/issues/8273">https://gitlab.alpinelinux.org/alpine/aports/issues/8273</a></p> <p>Release Date: 2017-12-07</p> <p>Fix Resolution: 1.0.2n</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2017-3738 (Medium) detected in opensslOpenSSL_1_1_0g - ## CVE-2017-3738 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>opensslOpenSSL_1_1_0g</b></p></summary> <p> <p>Akamai fork of openssl master.</p> <p>Library home page: <a href=https://github.com/akamai/openssl.git>https://github.com/akamai/openssl.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/Satheesh575555/Openssl_1_1_0_CVE-2021-23841/commit/4c0cda99bc44e45f32ea15e98aeb0a895ad1dd56">4c0cda99bc44e45f32ea15e98aeb0a895ad1dd56</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/crypto/bn/asm/rsaz-avx2.pl</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> There is an overflow bug in the AVX2 Montgomery multiplication procedure used in exponentiation with 1024-bit moduli. No EC algorithms are affected. Analysis suggests that attacks against RSA and DSA as a result of this defect would be very difficult to perform and are not believed likely. Attacks against DH1024 are considered just feasible, because most of the work necessary to deduce information about a private key may be performed offline. The amount of resources required for such an attack would be significant. However, for an attack on TLS to be meaningful, the server would have to share the DH1024 private key among multiple clients, which is no longer an option since CVE-2016-0701. This only affects processors that support the AVX2 but not ADX extensions like Intel Haswell (4th generation). Note: The impact from this issue is similar to CVE-2017-3736, CVE-2017-3732 and CVE-2015-3193. OpenSSL version 1.0.2-1.0.2m and 1.1.0-1.1.0g are affected. Fixed in OpenSSL 1.0.2n. Due to the low severity of this issue we are not issuing a new release of OpenSSL 1.1.0 at this time. The fix will be included in OpenSSL 1.1.0h when it becomes available. The fix is also available in commit e502cc86d in the OpenSSL git repository. <p>Publish Date: 2017-12-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-3738>CVE-2017-3738</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://gitlab.alpinelinux.org/alpine/aports/issues/8273">https://gitlab.alpinelinux.org/alpine/aports/issues/8273</a></p> <p>Release Date: 2017-12-07</p> <p>Fix Resolution: 1.0.2n</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in opensslopenssl cve medium severity vulnerability vulnerable library opensslopenssl akamai fork of openssl master library home page a href found in head commit a href found in base branch master vulnerable source files crypto bn asm rsaz pl vulnerability details there is an overflow bug in the montgomery multiplication procedure used in exponentiation with bit moduli no ec algorithms are affected analysis suggests that attacks against rsa and dsa as a result of this defect would be very difficult to perform and are not believed likely attacks against are considered just feasible because most of the work necessary to deduce information about a private key may be performed offline the amount of resources required for such an attack would be significant however for an attack on tls to be meaningful the server would have to share the private key among multiple clients which is no longer an option since cve this only affects processors that support the but not adx extensions like intel haswell generation note the impact from this issue is similar to cve cve and cve openssl version and are affected fixed in openssl due to the low severity of this issue we are not issuing a new release of openssl at this time the fix will be included in openssl when it becomes available the fix is also available in commit in the openssl git repository publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
40,798
16,546,026,188
IssuesEvent
2021-05-28 00:11:34
microsoft/vscode-cpptools
https://api.github.com/repos/microsoft/vscode-cpptools
closed
Wrong bitness for xtensa-esp32-elf-gcc
Feature: Configuration Language Service bug fixed (release pending) quick fix
I'm also having the same issue with GCC 8 and Xtensa (ESP32): ![Screenshot_20210217_105622](https://user-images.githubusercontent.com/424549/108187455-c85dbf80-710e-11eb-816d-f7605fef862d.png) This is obviously wrong, given that size_t IS unsigned int, and IntelliSense known that: ![Screenshot_20210217_105715](https://user-images.githubusercontent.com/424549/108187588-e9261500-710e-11eb-9d33-3efa78a08a74.png) Everything else works fine, except standard literals for strings (like `s` and `sv`), which are broken in the same way. Needless to say, this code compiles fine with GCC 8.4, triple is xtensa-esp32-elf, `sizeof(int) == 4`. The C/C++ extension I'm using is version 1.2.1. _Originally posted by @mcilloni in https://github.com/microsoft/vscode-cpptools/issues/6573#issuecomment-780442797_
1.0
Wrong bitness for xtensa-esp32-elf-gcc - I'm also having the same issue with GCC 8 and Xtensa (ESP32): ![Screenshot_20210217_105622](https://user-images.githubusercontent.com/424549/108187455-c85dbf80-710e-11eb-816d-f7605fef862d.png) This is obviously wrong, given that size_t IS unsigned int, and IntelliSense known that: ![Screenshot_20210217_105715](https://user-images.githubusercontent.com/424549/108187588-e9261500-710e-11eb-9d33-3efa78a08a74.png) Everything else works fine, except standard literals for strings (like `s` and `sv`), which are broken in the same way. Needless to say, this code compiles fine with GCC 8.4, triple is xtensa-esp32-elf, `sizeof(int) == 4`. The C/C++ extension I'm using is version 1.2.1. _Originally posted by @mcilloni in https://github.com/microsoft/vscode-cpptools/issues/6573#issuecomment-780442797_
non_defect
wrong bitness for xtensa elf gcc i m also having the same issue with gcc and xtensa this is obviously wrong given that size t is unsigned int and intellisense known that everything else works fine except standard literals for strings like s and sv which are broken in the same way needless to say this code compiles fine with gcc triple is xtensa elf sizeof int the c c extension i m using is version originally posted by mcilloni in
0
74,035
15,298,939,448
IssuesEvent
2021-02-24 10:19:47
rsoreq/kendo-ui-core
https://api.github.com/repos/rsoreq/kendo-ui-core
opened
CVE-2016-10531 (Medium) detected in marked-0.3.19.js
security vulnerability
## CVE-2016-10531 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>marked-0.3.19.js</b></p></summary> <p>A markdown parser built for speed</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/marked/0.3.19/marked.js">https://cdnjs.cloudflare.com/ajax/libs/marked/0.3.19/marked.js</a></p> <p>Path to dependency file: kendo-ui-core/docs/node_modules/marked/www/demo.html</p> <p>Path to vulnerable library: kendo-ui-core/docs/node_modules/marked/www/../lib/marked.js,kendo-ui-core/docs-aspnet/node_modules/marked/www/../lib/marked.js</p> <p> Dependency Hierarchy: - :x: **marked-0.3.19.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/rsoreq/kendo-ui-core/commit/62afbcdf79c4c7052417ecc86eb31bd6bc04e1ad">62afbcdf79c4c7052417ecc86eb31bd6bc04e1ad</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> marked is an application that is meant to parse and compile markdown. Due to the way that marked 0.3.5 and earlier parses input, specifically HTML entities, it's possible to bypass marked's content injection protection (`sanitize: true`) to inject a `javascript:` URL. This flaw exists because `&#xNNanything;` gets parsed to what it could and leaves the rest behind, resulting in just `anything;` being left. <p>Publish Date: 2018-05-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10531>CVE-2016-10531</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nodesecurity.io/advisories/101">https://nodesecurity.io/advisories/101</a></p> <p>Release Date: 2016-04-18</p> <p>Fix Resolution: Update to version 0.3.6 or later.</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"marked","packageVersion":"0.3.19","packageFilePaths":["/docs/node_modules/marked/www/demo.html","/docs-aspnet/node_modules/marked/www/demo.html"],"isTransitiveDependency":false,"dependencyTree":"marked:0.3.19","isMinimumFixVersionAvailable":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2016-10531","vulnerabilityDetails":"marked is an application that is meant to parse and compile markdown. Due to the way that marked 0.3.5 and earlier parses input, specifically HTML entities, it\u0027s possible to bypass marked\u0027s content injection protection (`sanitize: true`) to inject a `javascript:` URL. This flaw exists because `\u0026#xNNanything;` gets parsed to what it could and leaves the rest behind, resulting in just `anything;` being left.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10531","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2016-10531 (Medium) detected in marked-0.3.19.js - ## CVE-2016-10531 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>marked-0.3.19.js</b></p></summary> <p>A markdown parser built for speed</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/marked/0.3.19/marked.js">https://cdnjs.cloudflare.com/ajax/libs/marked/0.3.19/marked.js</a></p> <p>Path to dependency file: kendo-ui-core/docs/node_modules/marked/www/demo.html</p> <p>Path to vulnerable library: kendo-ui-core/docs/node_modules/marked/www/../lib/marked.js,kendo-ui-core/docs-aspnet/node_modules/marked/www/../lib/marked.js</p> <p> Dependency Hierarchy: - :x: **marked-0.3.19.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/rsoreq/kendo-ui-core/commit/62afbcdf79c4c7052417ecc86eb31bd6bc04e1ad">62afbcdf79c4c7052417ecc86eb31bd6bc04e1ad</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> marked is an application that is meant to parse and compile markdown. Due to the way that marked 0.3.5 and earlier parses input, specifically HTML entities, it's possible to bypass marked's content injection protection (`sanitize: true`) to inject a `javascript:` URL. This flaw exists because `&#xNNanything;` gets parsed to what it could and leaves the rest behind, resulting in just `anything;` being left. <p>Publish Date: 2018-05-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10531>CVE-2016-10531</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nodesecurity.io/advisories/101">https://nodesecurity.io/advisories/101</a></p> <p>Release Date: 2016-04-18</p> <p>Fix Resolution: Update to version 0.3.6 or later.</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"marked","packageVersion":"0.3.19","packageFilePaths":["/docs/node_modules/marked/www/demo.html","/docs-aspnet/node_modules/marked/www/demo.html"],"isTransitiveDependency":false,"dependencyTree":"marked:0.3.19","isMinimumFixVersionAvailable":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2016-10531","vulnerabilityDetails":"marked is an application that is meant to parse and compile markdown. Due to the way that marked 0.3.5 and earlier parses input, specifically HTML entities, it\u0027s possible to bypass marked\u0027s content injection protection (`sanitize: true`) to inject a `javascript:` URL. This flaw exists because `\u0026#xNNanything;` gets parsed to what it could and leaves the rest behind, resulting in just `anything;` being left.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10531","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_defect
cve medium detected in marked js cve medium severity vulnerability vulnerable library marked js a markdown parser built for speed library home page a href path to dependency file kendo ui core docs node modules marked www demo html path to vulnerable library kendo ui core docs node modules marked www lib marked js kendo ui core docs aspnet node modules marked www lib marked js dependency hierarchy x marked js vulnerable library found in head commit a href found in base branch master vulnerability details marked is an application that is meant to parse and compile markdown due to the way that marked and earlier parses input specifically html entities it s possible to bypass marked s content injection protection sanitize true to inject a javascript url this flaw exists because xnnanything gets parsed to what it could and leaves the rest behind resulting in just anything being left publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution update to version or later isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree marked isminimumfixversionavailable false basebranches vulnerabilityidentifier cve vulnerabilitydetails marked is an application that is meant to parse and compile markdown due to the way that marked and earlier parses input specifically html entities it possible to bypass marked content injection protection sanitize true to inject a javascript url this flaw exists because xnnanything gets parsed to what it could and leaves the rest behind resulting in just anything being left vulnerabilityurl
0
56,787
15,370,330,800
IssuesEvent
2021-03-02 08:40:05
MethodicalAcceleratorDesign/MAD-X
https://api.github.com/repos/MethodicalAcceleratorDesign/MAD-X
closed
Tilt + misalignment behaves strange in MAD-X PTC
PTC bug defect
It looks to be like the misalignment is applied in the tilted frame and is therefore moving the magnet in the wrong direction. I attach an example to illustrate: FASER : SEQUENCE, L = 10; b1:sbend, angle=138/180.0*pi, HGAP=0.0258,e1=0/180.0*pi, e2= 0/180.0*pi,FINT=0.5 ,TILT=1.0*pi, at=5,l=1; endsequence; beam; use, sequence=faser; eoption, add=true; select, flag=error, clear=true; select, flag=error, class=b1; ealign, dx= 0.001, dy= 0.00, ds=0.00,dphi=0.00, dtheta=0.00, dpsi=0.00; select, flag=twiss, column=name,s,x,px,y, py; twiss, betx=1, bety=1, x=-0.000, file="mytwiss.dat"; ptc_create_universe; ptc_create_layout, model=2, method=6,nst=10, exact; ptc_align; ptc_start, x=0.000; ptc_observe, place=b1; ptc_track, icase=5, element_by_element, dump, file="track.simple", ffile=1, onetable, maxaper={100.0, 100.0, 100.0, 100.0, 100.0, 100.0}; ptc_twiss, file="twiss.ptc.simple",betx=1,bety=1,betz=1; ptc_track_end; ptc_end;
1.0
Tilt + misalignment behaves strange in MAD-X PTC - It looks to be like the misalignment is applied in the tilted frame and is therefore moving the magnet in the wrong direction. I attach an example to illustrate: FASER : SEQUENCE, L = 10; b1:sbend, angle=138/180.0*pi, HGAP=0.0258,e1=0/180.0*pi, e2= 0/180.0*pi,FINT=0.5 ,TILT=1.0*pi, at=5,l=1; endsequence; beam; use, sequence=faser; eoption, add=true; select, flag=error, clear=true; select, flag=error, class=b1; ealign, dx= 0.001, dy= 0.00, ds=0.00,dphi=0.00, dtheta=0.00, dpsi=0.00; select, flag=twiss, column=name,s,x,px,y, py; twiss, betx=1, bety=1, x=-0.000, file="mytwiss.dat"; ptc_create_universe; ptc_create_layout, model=2, method=6,nst=10, exact; ptc_align; ptc_start, x=0.000; ptc_observe, place=b1; ptc_track, icase=5, element_by_element, dump, file="track.simple", ffile=1, onetable, maxaper={100.0, 100.0, 100.0, 100.0, 100.0, 100.0}; ptc_twiss, file="twiss.ptc.simple",betx=1,bety=1,betz=1; ptc_track_end; ptc_end;
defect
tilt misalignment behaves strange in mad x ptc it looks to be like the misalignment is applied in the tilted frame and is therefore moving the magnet in the wrong direction i attach an example to illustrate faser sequence l sbend angle pi hgap pi pi fint tilt pi at l endsequence beam use sequence faser eoption add true select flag error clear true select flag error class ealign dx dy ds dphi dtheta dpsi select flag twiss column name s x px y py twiss betx bety x file mytwiss dat ptc create universe ptc create layout model method nst exact ptc align ptc start x ptc observe place ptc track icase element by element dump file track simple ffile onetable maxaper ptc twiss file twiss ptc simple betx bety betz ptc track end ptc end
1
10,115
7,918,964,508
IssuesEvent
2018-07-04 14:58:25
maxdobeck/scheduler-frontend
https://api.github.com/repos/maxdobeck/scheduler-frontend
opened
Check for Good Session In Route Guard
bug security
User should be logged out or forced to log in again when session ends.
True
Check for Good Session In Route Guard - User should be logged out or forced to log in again when session ends.
non_defect
check for good session in route guard user should be logged out or forced to log in again when session ends
0
39,678
9,612,973,604
IssuesEvent
2019-05-13 10:09:01
contao/contao
https://api.github.com/repos/contao/contao
closed
Symlinks neu erstellen
defect
<a href="https://github.com/jommbee"><img src="https://avatars3.githubusercontent.com/u/4679889?v=4" align="left" width="42" height="42"></img></a> [Comment](https://github.com/contao/contao/issues/299#issuecomment-487871058) by @jommbee April 30th, 2019, 10:59 GMT Heute in der 4.4.39 nach Serverumzug gleicher Fehler. Nach manuellem Fix in der Zeile: https://github.com/contao/contao/blob/4.4/core-bundle/src/Resources/contao/library/Contao/Automator.php#L400 geht alles wunderbar.
1.0
Symlinks neu erstellen - <a href="https://github.com/jommbee"><img src="https://avatars3.githubusercontent.com/u/4679889?v=4" align="left" width="42" height="42"></img></a> [Comment](https://github.com/contao/contao/issues/299#issuecomment-487871058) by @jommbee April 30th, 2019, 10:59 GMT Heute in der 4.4.39 nach Serverumzug gleicher Fehler. Nach manuellem Fix in der Zeile: https://github.com/contao/contao/blob/4.4/core-bundle/src/Resources/contao/library/Contao/Automator.php#L400 geht alles wunderbar.
defect
symlinks neu erstellen by jommbee april gmt heute in der nach serverumzug gleicher fehler nach manuellem fix in der zeile geht alles wunderbar
1
240,180
26,254,331,125
IssuesEvent
2023-01-05 22:33:20
TreyM-WSS/terra-clinical
https://api.github.com/repos/TreyM-WSS/terra-clinical
opened
CVE-2021-23362 (Medium) detected in hosted-git-info-2.8.8.tgz
security vulnerability
## CVE-2021-23362 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hosted-git-info-2.8.8.tgz</b></p></summary> <p>Provides metadata and conversions from repository urls for Github, Bitbucket and Gitlab</p> <p>Library home page: <a href="https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz">https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/hosted-git-info/package.json</p> <p> Dependency Hierarchy: - check-installed-dependencies-1.0.1.tgz (Root Library) - meow-3.7.0.tgz - normalize-package-data-2.5.0.tgz - :x: **hosted-git-info-2.8.8.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/TreyM-WSS/terra-clinical/commit/0940364474cdf2812b34bb24824139d7e3f8060b">0940364474cdf2812b34bb24824139d7e3f8060b</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity. <p>Publish Date: 2021-03-23 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-23362>CVE-2021-23362</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-43f8-2h32-f4cj">https://github.com/advisories/GHSA-43f8-2h32-f4cj</a></p> <p>Release Date: 2021-03-23</p> <p>Fix Resolution: hosted-git-info - 2.8.9,3.0.8</p> </p> </details> <p></p>
True
CVE-2021-23362 (Medium) detected in hosted-git-info-2.8.8.tgz - ## CVE-2021-23362 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hosted-git-info-2.8.8.tgz</b></p></summary> <p>Provides metadata and conversions from repository urls for Github, Bitbucket and Gitlab</p> <p>Library home page: <a href="https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz">https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/hosted-git-info/package.json</p> <p> Dependency Hierarchy: - check-installed-dependencies-1.0.1.tgz (Root Library) - meow-3.7.0.tgz - normalize-package-data-2.5.0.tgz - :x: **hosted-git-info-2.8.8.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/TreyM-WSS/terra-clinical/commit/0940364474cdf2812b34bb24824139d7e3f8060b">0940364474cdf2812b34bb24824139d7e3f8060b</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity. <p>Publish Date: 2021-03-23 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-23362>CVE-2021-23362</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-43f8-2h32-f4cj">https://github.com/advisories/GHSA-43f8-2h32-f4cj</a></p> <p>Release Date: 2021-03-23</p> <p>Fix Resolution: hosted-git-info - 2.8.9,3.0.8</p> </p> </details> <p></p>
non_defect
cve medium detected in hosted git info tgz cve medium severity vulnerability vulnerable library hosted git info tgz provides metadata and conversions from repository urls for github bitbucket and gitlab library home page a href path to dependency file package json path to vulnerable library node modules hosted git info package json dependency hierarchy check installed dependencies tgz root library meow tgz normalize package data tgz x hosted git info tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package hosted git info before are vulnerable to regular expression denial of service redos via regular expression shortcutmatch in the fromurl function in index js the affected regular expression exhibits polynomial worst case time complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution hosted git info
0
15,626
2,866,060,794
IssuesEvent
2015-06-05 03:21:12
chaosben/theunknownones
https://api.github.com/repos/chaosben/theunknownones
closed
Fermeture of Google Code
auto-migrated Priority-Medium Type-Defect
``` Google Code will soon close. Have you thinked of another place to migrate, like GitHub (Using the "Export to GitHub" button)? ``` Original issue reported on code.google.com by `charlesm...@gmail.com` on 13 May 2015 at 10:15
1.0
Fermeture of Google Code - ``` Google Code will soon close. Have you thinked of another place to migrate, like GitHub (Using the "Export to GitHub" button)? ``` Original issue reported on code.google.com by `charlesm...@gmail.com` on 13 May 2015 at 10:15
defect
fermeture of google code google code will soon close have you thinked of another place to migrate like github using the export to github button original issue reported on code google com by charlesm gmail com on may at
1
74,821
25,345,831,442
IssuesEvent
2022-11-19 07:09:06
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
opened
fedora 37 upgrade fails when importing gpg keys
Type: Defect
<!-- Please fill out the following template, which will help other contributors address your issue. --> <!-- Thank you for reporting an issue. *IMPORTANT* - Please check our issue tracker before opening a new issue. Additional valuable information can be found in the OpenZFS documentation and mailing list archives. Please fill in as much of the template as possible. --> ### System information <!-- add version after "|" character --> Type | Version/Name --- | --- --> Distribution Name | fedora Distribution Version | 36 Kernel Version | 6.0.6-200.fc36 Architecture | x86-64 OpenZFS Version | zfs-2.1.6-1 <!-- Command to find OpenZFS version: zfs version Commands to find kernel version: uname -r # Linux freebsd-version -r # FreeBSD --> ### Describe the problem you're observing system-upgrade to --releasever=37 fails when dnf attempts to import gpg keys for zfs fc27 rpms ### Describe how to reproduce the problem sudo dnf -y system-upgrade download --allowerasing --refresh --releasever=37 dnf downloads zfs rpms from fc37 repo but fails when attempting to import gpg keys. I didn't see this problem when doing a system-upgrade to fc36. ### Include any warning/errors/backtraces from the system logs <!-- *IMPORTANT* - Please mark logs and text output from terminal commands or else Github will not display them correctly. An example is provided below. Example: --> ``` Fedora 37 - x86_64 1.6 MB/s | 1.6 kB 00:00 Importing GPG key 0x5323552A: Userid : "Fedora (37) <fedora-37-primary@fedoraproject.org>" Fingerprint: ACB5 EE4E 831C 74BB 7C16 8D27 F55A D3FB 5323 552A From : /etc/pki/rpm-gpg/RPM-GPG-KEY-fedora-37-x86_64 Key imported successfully ZFS on Linux for Fedora 37 1.6 MB/s | 1.7 kB 00:00 GPG key at file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux (0xF14AB620) is already installed The GPG keys listed for the "ZFS on Linux for Fedora 37" repository are already installed but they are not correct for this package. Check that the correct key URLs are configured for this repository.. Failing package is: libnvpair3-2.1.6-2.fc37.x86_64 GPG Keys are configured as: file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux Public key for libuutil3-2.1.6-2.fc37.x86_64.rpm is not installed. Failing package is: libuutil3-2.1.6-2.fc37.x86_64 GPG Keys are configured as: file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux Public key for libzfs5-2.1.6-2.fc37.x86_64.rpm is not installed. Failing package is: libzfs5-2.1.6-2.fc37.x86_64 GPG Keys are configured as: file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux Public key for libzpool5-2.1.6-2.fc37.x86_64.rpm is not installed. Failing package is: libzpool5-2.1.6-2.fc37.x86_64 GPG Keys are configured as: file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux Public key for zfs-2.1.6-2.fc37.x86_64.rpm is not installed. Failing package is: zfs-2.1.6-2.fc37.x86_64 GPG Keys are configured as: file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux Public key for zfs-dkms-2.1.6-2.fc37.noarch.rpm is not installed. Failing package is: zfs-dkms-2.1.6-2.fc37.noarch GPG Keys are configured as: file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux Public key for zfs-dracut-2.1.6-2.fc37.noarch.rpm is not installed. Failing package is: zfs-dracut-2.1.6-2.fc37.noarch GPG Keys are configured as: file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux The downloaded packages were saved in cache until the next successful transaction. You can remove cached packages by executing 'dnf clean packages'. Error: GPG check FAILED ``` file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux
1.0
fedora 37 upgrade fails when importing gpg keys - <!-- Please fill out the following template, which will help other contributors address your issue. --> <!-- Thank you for reporting an issue. *IMPORTANT* - Please check our issue tracker before opening a new issue. Additional valuable information can be found in the OpenZFS documentation and mailing list archives. Please fill in as much of the template as possible. --> ### System information <!-- add version after "|" character --> Type | Version/Name --- | --- --> Distribution Name | fedora Distribution Version | 36 Kernel Version | 6.0.6-200.fc36 Architecture | x86-64 OpenZFS Version | zfs-2.1.6-1 <!-- Command to find OpenZFS version: zfs version Commands to find kernel version: uname -r # Linux freebsd-version -r # FreeBSD --> ### Describe the problem you're observing system-upgrade to --releasever=37 fails when dnf attempts to import gpg keys for zfs fc27 rpms ### Describe how to reproduce the problem sudo dnf -y system-upgrade download --allowerasing --refresh --releasever=37 dnf downloads zfs rpms from fc37 repo but fails when attempting to import gpg keys. I didn't see this problem when doing a system-upgrade to fc36. ### Include any warning/errors/backtraces from the system logs <!-- *IMPORTANT* - Please mark logs and text output from terminal commands or else Github will not display them correctly. An example is provided below. Example: --> ``` Fedora 37 - x86_64 1.6 MB/s | 1.6 kB 00:00 Importing GPG key 0x5323552A: Userid : "Fedora (37) <fedora-37-primary@fedoraproject.org>" Fingerprint: ACB5 EE4E 831C 74BB 7C16 8D27 F55A D3FB 5323 552A From : /etc/pki/rpm-gpg/RPM-GPG-KEY-fedora-37-x86_64 Key imported successfully ZFS on Linux for Fedora 37 1.6 MB/s | 1.7 kB 00:00 GPG key at file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux (0xF14AB620) is already installed The GPG keys listed for the "ZFS on Linux for Fedora 37" repository are already installed but they are not correct for this package. Check that the correct key URLs are configured for this repository.. Failing package is: libnvpair3-2.1.6-2.fc37.x86_64 GPG Keys are configured as: file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux Public key for libuutil3-2.1.6-2.fc37.x86_64.rpm is not installed. Failing package is: libuutil3-2.1.6-2.fc37.x86_64 GPG Keys are configured as: file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux Public key for libzfs5-2.1.6-2.fc37.x86_64.rpm is not installed. Failing package is: libzfs5-2.1.6-2.fc37.x86_64 GPG Keys are configured as: file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux Public key for libzpool5-2.1.6-2.fc37.x86_64.rpm is not installed. Failing package is: libzpool5-2.1.6-2.fc37.x86_64 GPG Keys are configured as: file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux Public key for zfs-2.1.6-2.fc37.x86_64.rpm is not installed. Failing package is: zfs-2.1.6-2.fc37.x86_64 GPG Keys are configured as: file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux Public key for zfs-dkms-2.1.6-2.fc37.noarch.rpm is not installed. Failing package is: zfs-dkms-2.1.6-2.fc37.noarch GPG Keys are configured as: file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux Public key for zfs-dracut-2.1.6-2.fc37.noarch.rpm is not installed. Failing package is: zfs-dracut-2.1.6-2.fc37.noarch GPG Keys are configured as: file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux The downloaded packages were saved in cache until the next successful transaction. You can remove cached packages by executing 'dnf clean packages'. Error: GPG check FAILED ``` file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux
defect
fedora upgrade fails when importing gpg keys thank you for reporting an issue important please check our issue tracker before opening a new issue additional valuable information can be found in the openzfs documentation and mailing list archives please fill in as much of the template as possible system information type version name distribution name fedora distribution version kernel version architecture openzfs version zfs command to find openzfs version zfs version commands to find kernel version uname r linux freebsd version r freebsd describe the problem you re observing system upgrade to releasever fails when dnf attempts to import gpg keys for zfs rpms describe how to reproduce the problem sudo dnf y system upgrade download allowerasing refresh releasever dnf downloads zfs rpms from repo but fails when attempting to import gpg keys i didn t see this problem when doing a system upgrade to include any warning errors backtraces from the system logs important please mark logs and text output from terminal commands or else github will not display them correctly an example is provided below example fedora mb s kb importing gpg key userid fedora fingerprint from etc pki rpm gpg rpm gpg key fedora key imported successfully zfs on linux for fedora mb s kb gpg key at file etc pki rpm gpg rpm gpg key zfsonlinux is already installed the gpg keys listed for the zfs on linux for fedora repository are already installed but they are not correct for this package check that the correct key urls are configured for this repository failing package is gpg keys are configured as file etc pki rpm gpg rpm gpg key zfsonlinux public key for rpm is not installed failing package is gpg keys are configured as file etc pki rpm gpg rpm gpg key zfsonlinux public key for rpm is not installed failing package is gpg keys are configured as file etc pki rpm gpg rpm gpg key zfsonlinux public key for rpm is not installed failing package is gpg keys are configured as file etc pki rpm gpg rpm gpg key zfsonlinux public key for zfs rpm is not installed failing package is zfs gpg keys are configured as file etc pki rpm gpg rpm gpg key zfsonlinux public key for zfs dkms noarch rpm is not installed failing package is zfs dkms noarch gpg keys are configured as file etc pki rpm gpg rpm gpg key zfsonlinux public key for zfs dracut noarch rpm is not installed failing package is zfs dracut noarch gpg keys are configured as file etc pki rpm gpg rpm gpg key zfsonlinux the downloaded packages were saved in cache until the next successful transaction you can remove cached packages by executing dnf clean packages error gpg check failed file etc pki rpm gpg rpm gpg key zfsonlinux
1