Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 757 | labels stringlengths 4 664 | body stringlengths 3 261k | index stringclasses 10 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 232k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
19,878 | 4,454,497,887 | IssuesEvent | 2016-08-23 01:08:48 | evhub/coconut | https://api.github.com/repos/evhub/coconut | closed | Improve iterator matching documentation | documentation | Because Python iterators aren't real lazy lists, you have to be very careful about accidentally consuming them when performing checks, which makes iterator matching rather confusing. | 1.0 | Improve iterator matching documentation - Because Python iterators aren't real lazy lists, you have to be very careful about accidentally consuming them when performing checks, which makes iterator matching rather confusing. | non_defect | improve iterator matching documentation because python iterators aren t real lazy lists you have to be very careful about accidentally consuming them when performing checks which makes iterator matching rather confusing | 0 |
172,637 | 13,325,483,657 | IssuesEvent | 2020-08-27 10:02:00 | gambitph/Stackable | https://api.github.com/repos/gambitph/Stackable | opened | Social Snap Plugin "Pin It" feature is adding some custom CSS code to the image of the Testimonial block making the image small | [block] testimonial bug | Social Snap Plugin "Pin It" feature adds a Pinterest button overall images, making it small.

Here's the code to fix the issue:
`.ss-on-media-image-wrap.ugb-img--shape {
display: block !important;
}`
| 1.0 | Social Snap Plugin "Pin It" feature is adding some custom CSS code to the image of the Testimonial block making the image small - Social Snap Plugin "Pin It" feature adds a Pinterest button overall images, making it small.

Here's the code to fix the issue:
`.ss-on-media-image-wrap.ugb-img--shape {
display: block !important;
}`
| non_defect | social snap plugin pin it feature is adding some custom css code to the image of the testimonial block making the image small social snap plugin pin it feature adds a pinterest button overall images making it small here s the code to fix the issue ss on media image wrap ugb img shape display block important | 0 |
81,023 | 30,662,370,985 | IssuesEvent | 2023-07-25 15:47:22 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | Hazelcast documentation incorrectly states that "local" option is available for ItemListener [HZ-2777] | Type: Defect Team: Core Type: Cleanup Source: Community Module: IQueue Module: ISet Module: ReplicatedMap Module: IList Type: Documentation to-jira | <!--
Thanks for reporting your issue. Please share with us the following information, to help us resolve your issue quickly and efficiently.
-->
**Describe the bug**
The documentation states:
`item-listeners: Lets you add listeners (listener classes) for the list items. ... You can set **local** to true if you want to listen to the items on the local member.`
**Expected behavior**
The **local** option is not available for IQueue, ISet, IList or ReplicatedMap ItemListener.
**To Reproduce**
instance.getQueue("someQueue")
.addItemListener(new MyItemListener<>(), true);
No option to specify if the listener should only run on local member that hosts the queue.
| 1.0 | Hazelcast documentation incorrectly states that "local" option is available for ItemListener [HZ-2777] - <!--
Thanks for reporting your issue. Please share with us the following information, to help us resolve your issue quickly and efficiently.
-->
**Describe the bug**
The documentation states:
`item-listeners: Lets you add listeners (listener classes) for the list items. ... You can set **local** to true if you want to listen to the items on the local member.`
**Expected behavior**
The **local** option is not available for IQueue, ISet, IList or ReplicatedMap ItemListener.
**To Reproduce**
instance.getQueue("someQueue")
.addItemListener(new MyItemListener<>(), true);
No option to specify if the listener should only run on local member that hosts the queue.
| defect | hazelcast documentation incorrectly states that local option is available for itemlistener thanks for reporting your issue please share with us the following information to help us resolve your issue quickly and efficiently describe the bug the documentation states item listeners lets you add listeners listener classes for the list items you can set local to true if you want to listen to the items on the local member expected behavior the local option is not available for iqueue iset ilist or replicatedmap itemlistener to reproduce instance getqueue somequeue additemlistener new myitemlistener true no option to specify if the listener should only run on local member that hosts the queue | 1 |
35,338 | 7,702,684,008 | IssuesEvent | 2018-05-21 04:13:50 | gperftools/gperftools | https://api.github.com/repos/gperftools/gperftools | closed | Option to disable writing tcmalloc larg allocation messages to stderr | Priority-Medium Status-WontFix Type-Defect | Originally reported on Google Code with ID 357
```
tcmalloc will print messages such as the following to stderr:
tcmalloc: large alloc 1176436736 bytes == 0x3177a000 @ 0x7f909f9fce6c 0x401ef9 0x40b931
0x44f1da 0x44be5f 0x4502c3 0x7f909ed9bc4d 0x401db9
Looking at the code, the output to stderr looks built in. Is there a mechanism for
turning this off, as not to contaminate stderr unless there is actually an error?
```
Reported by `nilshomer` on 2011-07-29 19:43:15
| 1.0 | Option to disable writing tcmalloc larg allocation messages to stderr - Originally reported on Google Code with ID 357
```
tcmalloc will print messages such as the following to stderr:
tcmalloc: large alloc 1176436736 bytes == 0x3177a000 @ 0x7f909f9fce6c 0x401ef9 0x40b931
0x44f1da 0x44be5f 0x4502c3 0x7f909ed9bc4d 0x401db9
Looking at the code, the output to stderr looks built in. Is there a mechanism for
turning this off, as not to contaminate stderr unless there is actually an error?
```
Reported by `nilshomer` on 2011-07-29 19:43:15
| defect | option to disable writing tcmalloc larg allocation messages to stderr originally reported on google code with id tcmalloc will print messages such as the following to stderr tcmalloc large alloc bytes looking at the code the output to stderr looks built in is there a mechanism for turning this off as not to contaminate stderr unless there is actually an error reported by nilshomer on | 1 |
49,010 | 13,185,192,482 | IssuesEvent | 2020-08-12 20:54:28 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | opened | check for moved/copied tools (Trac #598) | Incomplete Migration Migrated from Trac cmake defect | <details>
<summary><em>Migrated from https://code.icecube.wisc.edu/ticket/598
, reported by troy and owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2012-06-17T19:55:06",
"description": "not clear when to check, perhaps in the 'port' command and when running cmake.\n\n{{{\n\n14:15 <molivo> it says \"qmake found at bla bla bla\" then two warnings QT_QMAKE_EXECUTABLE reported QT_INSTALL_LIBS ...\n14:15 <molivo> wait a sec i am a bit messed up :-)\n14:20 <molivo> -- qmake found at /rest/tools/qt-4.6.0/bin/qmake\n14:20 <molivo> Warning: QT_QMAKE_EXECUTABLE reported QT_INSTALL_LIBS as /rest/icecube/tools/qt-4.6.0/lib\n14:20 <molivo> Warning: /rest/icecube/tools/qt-4.6.0/lib does NOT exist, Qt must NOT be installed correctly.\n14:20 <molivo> this is what I get\n14:21 <straszhm> hm, and that directory does exist... you said it contains libQtOpenGL.so\n14:21 <straszhm> (?)\n14:21 <molivo> yes it does\n14:22 <straszhm> random try: rm your CMakeCache.txt and run cmake again\n14:22 <molivo> ok\n14:22 <straszhm> hrm\n14:22 <straszhm> wait,\n14:22 <straszhm> -- qmake found at /rest/tools/qt-4.6.0/bin/qmake\n14:22 <molivo> yes\n14:22 <straszhm> reported QT_INSTALL_LIBS as /rest/icecube/tools/qt-4.6.0/lib\n14:22 <straszhm> those are two different directories\n14:23 <molivo> yes...\n14:23 <straszhm> did you move the toolset after you built it?\n14:23 <molivo> well, I did not. I copied from jens' pc\n14:23 <molivo> maybe he did\n}}}",
"reporter": "troy",
"cc": "",
"resolution": "wont or cant fix",
"_ts": "1339962906000000",
"component": "cmake",
"summary": "check for moved/copied tools",
"priority": "normal",
"keywords": "",
"time": "2010-02-23T19:26:18",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
| 1.0 | check for moved/copied tools (Trac #598) - <details>
<summary><em>Migrated from https://code.icecube.wisc.edu/ticket/598
, reported by troy and owned by nega</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2012-06-17T19:55:06",
"description": "not clear when to check, perhaps in the 'port' command and when running cmake.\n\n{{{\n\n14:15 <molivo> it says \"qmake found at bla bla bla\" then two warnings QT_QMAKE_EXECUTABLE reported QT_INSTALL_LIBS ...\n14:15 <molivo> wait a sec i am a bit messed up :-)\n14:20 <molivo> -- qmake found at /rest/tools/qt-4.6.0/bin/qmake\n14:20 <molivo> Warning: QT_QMAKE_EXECUTABLE reported QT_INSTALL_LIBS as /rest/icecube/tools/qt-4.6.0/lib\n14:20 <molivo> Warning: /rest/icecube/tools/qt-4.6.0/lib does NOT exist, Qt must NOT be installed correctly.\n14:20 <molivo> this is what I get\n14:21 <straszhm> hm, and that directory does exist... you said it contains libQtOpenGL.so\n14:21 <straszhm> (?)\n14:21 <molivo> yes it does\n14:22 <straszhm> random try: rm your CMakeCache.txt and run cmake again\n14:22 <molivo> ok\n14:22 <straszhm> hrm\n14:22 <straszhm> wait,\n14:22 <straszhm> -- qmake found at /rest/tools/qt-4.6.0/bin/qmake\n14:22 <molivo> yes\n14:22 <straszhm> reported QT_INSTALL_LIBS as /rest/icecube/tools/qt-4.6.0/lib\n14:22 <straszhm> those are two different directories\n14:23 <molivo> yes...\n14:23 <straszhm> did you move the toolset after you built it?\n14:23 <molivo> well, I did not. I copied from jens' pc\n14:23 <molivo> maybe he did\n}}}",
"reporter": "troy",
"cc": "",
"resolution": "wont or cant fix",
"_ts": "1339962906000000",
"component": "cmake",
"summary": "check for moved/copied tools",
"priority": "normal",
"keywords": "",
"time": "2010-02-23T19:26:18",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
</p>
</details>
| defect | check for moved copied tools trac migrated from reported by troy and owned by nega json status closed changetime description not clear when to check perhaps in the port command and when running cmake n n n it says qmake found at bla bla bla then two warnings qt qmake executable reported qt install libs wait a sec i am a bit messed up qmake found at rest tools qt bin qmake warning qt qmake executable reported qt install libs as rest icecube tools qt lib warning rest icecube tools qt lib does not exist qt must not be installed correctly this is what i get hm and that directory does exist you said it contains libqtopengl so yes it does random try rm your cmakecache txt and run cmake again ok hrm wait qmake found at rest tools qt bin qmake yes reported qt install libs as rest icecube tools qt lib those are two different directories yes did you move the toolset after you built it well i did not i copied from jens pc maybe he did n reporter troy cc resolution wont or cant fix ts component cmake summary check for moved copied tools priority normal keywords time milestone owner nega type defect | 1 |
39,273 | 9,368,595,978 | IssuesEvent | 2019-04-03 09:03:52 | Automattic/wp-calypso | https://api.github.com/repos/Automattic/wp-calypso | closed | SDK: Build source outside Calypso repo | [Project] SDK [Type] Defect | The SDK should be able to build from source outside of Calypso. Currently, it cannot.
#### Steps to reproduce
Try to build some source that isn't inside of Calypso:
```sh
npm run sdk generic /some/path/to/some/file.js output/file.js
# Module not found: Error: Can't resolve '@babel/runtime/helpers/typeof' in …
```
It fails
cc: @Automattic/calypso-sdk | 1.0 | SDK: Build source outside Calypso repo - The SDK should be able to build from source outside of Calypso. Currently, it cannot.
#### Steps to reproduce
Try to build some source that isn't inside of Calypso:
```sh
npm run sdk generic /some/path/to/some/file.js output/file.js
# Module not found: Error: Can't resolve '@babel/runtime/helpers/typeof' in …
```
It fails
cc: @Automattic/calypso-sdk | defect | sdk build source outside calypso repo the sdk should be able to build from source outside of calypso currently it cannot steps to reproduce try to build some source that isn t inside of calypso sh npm run sdk generic some path to some file js output file js module not found error can t resolve babel runtime helpers typeof in … it fails cc automattic calypso sdk | 1 |
431,282 | 12,476,697,067 | IssuesEvent | 2020-05-29 13:54:40 | medic/cht-core | https://api.github.com/repos/medic/cht-core | closed | Support resolving tasks concerning unknown contacts | Priority: 1 - High Type: Bug |
**Describe the bug**
We can no longer rely on `action.modifyContent` to inject contact information into reports generated by task actions.
This causes tasks that are about unknown contacts to generate reports that are unattached (are missing their subject).
This workflow is used where supervisors with replication depth receive `needs_signoff` reports about patients they can't see.
**To Reproduce**
1. Create a supervisor with replication_depth low enough that they don't see patients.
2. Create a chw under this supervisor.
3. Configure a form with needs_signoff.
4. Configure a task for the supervisor triggered by the `needs_signoff` form. Configure this to be resolved if the task action is executed.
5. Submit the `needs_signoff` form as the CHW.
6. Sync as the supervisor and check if you have generated the task.
7. Complete the task.
8. Your task does not resolve (it's still in the list) and inspecting the resulting report, it's missing all patient information.
**Expected behavior**
The resulting report should have the same patient information as the generating report.
**Environment**
- Instance: local
- App: webapp
- Version: 3.9
**Additional context**
When written to disk, tasks are minified and they only retain a "forId" property representing `action.contact._id`, while `action.contact` is deleted.
https://github.com/medic/cht-core/blob/master/shared-libs/rules-engine/src/transform-task-emission-to-doc.js#L77
When rehydrating the task, we rely on the `forId` contact to exist in order to repopulate `action.contact`: https://github.com/medic/cht-core/blob/master/webapp/src/js/controllers/tasks-content.js#L109
We should either:
a) keep `action.contact` when the contact is unknown - this allows more information besides a uuid to be injected into the resulting report
b) use a stub when rehydrating and contact is unknown - this means that the resulting report will only have the uuid available and no other information. | 1.0 | Support resolving tasks concerning unknown contacts -
**Describe the bug**
We can no longer rely on `action.modifyContent` to inject contact information into reports generated by task actions.
This causes tasks that are about unknown contacts to generate reports that are unattached (are missing their subject).
This workflow is used where supervisors with replication depth receive `needs_signoff` reports about patients they can't see.
**To Reproduce**
1. Create a supervisor with replication_depth low enough that they don't see patients.
2. Create a chw under this supervisor.
3. Configure a form with needs_signoff.
4. Configure a task for the supervisor triggered by the `needs_signoff` form. Configure this to be resolved if the task action is executed.
5. Submit the `needs_signoff` form as the CHW.
6. Sync as the supervisor and check if you have generated the task.
7. Complete the task.
8. Your task does not resolve (it's still in the list) and inspecting the resulting report, it's missing all patient information.
**Expected behavior**
The resulting report should have the same patient information as the generating report.
**Environment**
- Instance: local
- App: webapp
- Version: 3.9
**Additional context**
When written to disk, tasks are minified and they only retain a "forId" property representing `action.contact._id`, while `action.contact` is deleted.
https://github.com/medic/cht-core/blob/master/shared-libs/rules-engine/src/transform-task-emission-to-doc.js#L77
When rehydrating the task, we rely on the `forId` contact to exist in order to repopulate `action.contact`: https://github.com/medic/cht-core/blob/master/webapp/src/js/controllers/tasks-content.js#L109
We should either:
a) keep `action.contact` when the contact is unknown - this allows more information besides a uuid to be injected into the resulting report
b) use a stub when rehydrating and contact is unknown - this means that the resulting report will only have the uuid available and no other information. | non_defect | support resolving tasks concerning unknown contacts describe the bug we can no longer rely on action modifycontent to inject contact information into reports generated by task actions this causes tasks that are about unknown contacts to generate reports that are unattached are missing their subject this workflow is used where supervisors with replication depth receive needs signoff reports about patients they can t see to reproduce create a supervisor with replication depth low enough that they don t see patients create a chw under this supervisor configure a form with needs signoff configure a task for the supervisor triggered by the needs signoff form configure this to be resolved if the task action is executed submit the needs signoff form as the chw sync as the supervisor and check if you have generated the task complete the task your task does not resolve it s still in the list and inspecting the resulting report it s missing all patient information expected behavior the resulting report should have the same patient information as the generating report environment instance local app webapp version additional context when written to disk tasks are minified and they only retain a forid property representing action contact id while action contact is deleted when rehydrating the task we rely on the forid contact to exist in order to repopulate action contact we should either a keep action contact when the contact is unknown this allows more information besides a uuid to be injected into the resulting report b use a stub when rehydrating and contact is unknown this means that the resulting report will only have the uuid available and no other information | 0 |
75,656 | 25,978,379,317 | IssuesEvent | 2022-12-19 16:36:51 | department-of-veterans-affairs/va.gov-cms | https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms | closed | FE: VAMC System Banner Alert with Situation Updates do not work for Lovell systems | Defect VA.gov frontend ⭐️ Facilities Needs refining | ## Describe the defect
This issue will need to be refined as work in #11726 is completed. For now this is a placeholder. We simply want to be sure that VAMC System Banner Alert with Situation Updates nodes are displayed properly for Lovell systems...
## To Reproduce
Steps to reproduce the behavior:
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. See error
## AC / Expected behavior
A clear and concise description of what you expected to happen.
## Screenshots
If applicable, add screenshots to help explain your problem.
## Additional context
Add any other context about the problem here. Reach out to the Product Managers to determine if it should be escalated as critical (prevents users from accomplishing their work with no known workaround and needs to be addressed within 2 business days).
## Desktop (please complete the following information if relevant, or delete)
- OS: [e.g. iOS]
- Browser [e.g. chrome, safari]
- Version [e.g. 22]
## Labels
(You can delete this section once it's complete)
- [x] Issue type (red) (defaults to "Defect")
- [ ] CMS subsystem (green)
- [ ] CMS practice area (blue)
- [x] CMS workstream (orange) (not needed for bug tickets)
- [ ] CMS-supported product (black)
### CMS Team
Please check the team(s) that will do this work.
- [ ] `Program`
- [ ] `Platform CMS Team`
- [ ] `Sitewide Crew`
- [ ] `⭐️ Sitewide CMS`
- [ ] `⭐️ Public Websites`
- [ ] `⭐️ Facilities`
- [ ] `⭐️ User support`
| 1.0 | FE: VAMC System Banner Alert with Situation Updates do not work for Lovell systems - ## Describe the defect
This issue will need to be refined as work in #11726 is completed. For now this is a placeholder. We simply want to be sure that VAMC System Banner Alert with Situation Updates nodes are displayed properly for Lovell systems...
## To Reproduce
Steps to reproduce the behavior:
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. See error
## AC / Expected behavior
A clear and concise description of what you expected to happen.
## Screenshots
If applicable, add screenshots to help explain your problem.
## Additional context
Add any other context about the problem here. Reach out to the Product Managers to determine if it should be escalated as critical (prevents users from accomplishing their work with no known workaround and needs to be addressed within 2 business days).
## Desktop (please complete the following information if relevant, or delete)
- OS: [e.g. iOS]
- Browser [e.g. chrome, safari]
- Version [e.g. 22]
## Labels
(You can delete this section once it's complete)
- [x] Issue type (red) (defaults to "Defect")
- [ ] CMS subsystem (green)
- [ ] CMS practice area (blue)
- [x] CMS workstream (orange) (not needed for bug tickets)
- [ ] CMS-supported product (black)
### CMS Team
Please check the team(s) that will do this work.
- [ ] `Program`
- [ ] `Platform CMS Team`
- [ ] `Sitewide Crew`
- [ ] `⭐️ Sitewide CMS`
- [ ] `⭐️ Public Websites`
- [ ] `⭐️ Facilities`
- [ ] `⭐️ User support`
| defect | fe vamc system banner alert with situation updates do not work for lovell systems describe the defect this issue will need to be refined as work in is completed for now this is a placeholder we simply want to be sure that vamc system banner alert with situation updates nodes are displayed properly for lovell systems to reproduce steps to reproduce the behavior go to click on scroll down to see error ac expected behavior a clear and concise description of what you expected to happen screenshots if applicable add screenshots to help explain your problem additional context add any other context about the problem here reach out to the product managers to determine if it should be escalated as critical prevents users from accomplishing their work with no known workaround and needs to be addressed within business days desktop please complete the following information if relevant or delete os browser version labels you can delete this section once it s complete issue type red defaults to defect cms subsystem green cms practice area blue cms workstream orange not needed for bug tickets cms supported product black cms team please check the team s that will do this work program platform cms team sitewide crew ⭐️ sitewide cms ⭐️ public websites ⭐️ facilities ⭐️ user support | 1 |
177,457 | 28,494,894,530 | IssuesEvent | 2023-04-18 13:37:14 | rancher/dashboard | https://api.github.com/repos/rancher/dashboard | opened | Kubernetes RBAC competitive analysis | kind/design | We want to understand how other actors deal with the particularities and complexities of Kubernetes RBAC.
The initial idea is to compare Rancher's approach with:
- K8s Dashboard
- Lens
- Octant
| 1.0 | Kubernetes RBAC competitive analysis - We want to understand how other actors deal with the particularities and complexities of Kubernetes RBAC.
The initial idea is to compare Rancher's approach with:
- K8s Dashboard
- Lens
- Octant
| non_defect | kubernetes rbac competitive analysis we want to understand how other actors deal with the particularities and complexities of kubernetes rbac the initial idea is to compare rancher s approach with dashboard lens octant | 0 |
493,417 | 14,231,702,816 | IssuesEvent | 2020-11-18 09:55:51 | kubermatic/kubermatic | https://api.github.com/repos/kubermatic/kubermatic | closed | Support MachineDeployment specific kubelet config | customer-request lifecycle/rotten priority/low team/lifecycle | **User Story**
The kubelet configuration is currently applied as a ConfigMap from the addons image. Customer wants to change the API request limits per cluster or even MachineDeployment. E.g., different limits for small and large nodes.
**Acceptance criteria**
Kubelet configuration should be adaptable per MachineDeployment | 1.0 | Support MachineDeployment specific kubelet config - **User Story**
The kubelet configuration is currently applied as a ConfigMap from the addons image. Customer wants to change the API request limits per cluster or even MachineDeployment. E.g., different limits for small and large nodes.
**Acceptance criteria**
Kubelet configuration should be adaptable per MachineDeployment | non_defect | support machinedeployment specific kubelet config user story the kubelet configuration is currently applied as a configmap from the addons image customer wants to change the api request limits per cluster or even machinedeployment e g different limits for small and large nodes acceptance criteria kubelet configuration should be adaptable per machinedeployment | 0 |
542,929 | 15,874,406,843 | IssuesEvent | 2021-04-09 04:57:34 | remnoteio/remnote-issues | https://api.github.com/repos/remnoteio/remnote-issues | closed | No. of Tagged Rems does not appear or appear incorrectly at bottom portal results | priority=2 | Although there is indication of tagged rems existing in the bottom portal, upon expanding the portal there would be no results shown.
In other instances, an incorrect count of Tagged Rems is reflected.
---

---

---

---

| 1.0 | No. of Tagged Rems does not appear or appear incorrectly at bottom portal results - Although there is indication of tagged rems existing in the bottom portal, upon expanding the portal there would be no results shown.
In other instances, an incorrect count of Tagged Rems is reflected.
---

---

---

---

| non_defect | no of tagged rems does not appear or appear incorrectly at bottom portal results although there is indication of tagged rems existing in the bottom portal upon expanding the portal there would be no results shown in other instances an incorrect count of tagged rems is reflected | 0 |
78,768 | 27,751,928,124 | IssuesEvent | 2023-03-15 21:31:41 | jurgendl/hql-builder | https://api.github.com/repos/jurgendl/hql-builder | closed | remove HikariCP, fallback to previous connection-pool | bug Type-Defect Priority-Critical no-issue-activity | remove HikariCP, fallback to previous connection-pool
HikariCP seems to hang
| 1.0 | remove HikariCP, fallback to previous connection-pool - remove HikariCP, fallback to previous connection-pool
HikariCP seems to hang
| defect | remove hikaricp fallback to previous connection pool remove hikaricp fallback to previous connection pool hikaricp seems to hang | 1 |
67,673 | 8,170,283,302 | IssuesEvent | 2018-08-27 07:04:35 | otavanopisto/muikku | https://api.github.com/repos/otavanopisto/muikku | closed | Audio Component | MATERIALS REDESIGN2017 enhancement | _From @onzag on August 27, 2018 6:3_
null
_Copied from original issue: otavanopisto/board#1031_ | 1.0 | Audio Component - _From @onzag on August 27, 2018 6:3_
null
_Copied from original issue: otavanopisto/board#1031_ | non_defect | audio component from onzag on august null copied from original issue otavanopisto board | 0 |
75,397 | 25,810,854,823 | IssuesEvent | 2022-12-11 20:56:38 | DependencyTrack/dependency-track | https://api.github.com/repos/DependencyTrack/dependency-track | closed | Dependency track takes more than 30 minutes to analyze and Dependency tracker jenkins aborts due to polling time limti. | defect | The defect may already be reported! Please search for the defect before creating one.
### Current Behavior:
Jenkins Dependecy tarck jobs fails since server takes more 30 minutes analyze the project.
### Steps to Reproduce:
Run jenkins Dependecy track job in Jenkins pipeline for java project which has more than 1000 components(These are java modules)
### Expected Behavior:
Jenkins PipleLine should not fail.
### Environment:
- Dependency-Track Version: 'latest'
- Distribution: [ Docker ]
- BOM Format & Version: "bomFormat": "CycloneDX",
"specVersion": "1.4",
"version": 1,
- Database Server: PostgreSQL
- Browser: Chrome
### Additional Details:
22-06-28 06:37:51,463 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed portfolio metrics update
2022-06-28 06:37:56,340 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed metrics update for project: cd6feaf1-c6df-4c15-a7b4-f118151a6605
2022-06-28 06:52:29,046 [] INFO [org.dependencytrack.tasks.BomUploadProcessingTask] Processing CycloneDX BOM uploaded to project: cd6feaf1-c6df-4c15-a7b4-f118151a6605
2022-06-28 06:56:59,202 [] INFO [org.dependencytrack.tasks.BomUploadProcessingTask] Processing CycloneDX dependency graph for project: cd6feaf1-c6df-4c15-a7b4-f118151a6605
2022-06-28 07:04:14,351 [] INFO [org.dependencytrack.tasks.BomUploadProcessingTask] Processed 18477 components and 0 services uploaded to project cd6feaf1-c6df-4c15-a7b4-f118151a6605
2022-06-28 07:22:08,107 [] INFO [org.dependencytrack.tasks.scanners.InternalAnalysisTask] Starting internal analysis task
2022-06-28 07:22:42,573 [] INFO [org.dependencytrack.tasks.scanners.InternalAnalysisTask] Internal analysis complete
2022-06-28 07:22:42,579 [] WARN [org.dependencytrack.tasks.scanners.OssIndexAnalysisTask] An API username or token has not been specified for use with OSS Index. Using anonymous access
2022-06-28 07:22:42,579 [] INFO [org.dependencytrack.tasks.scanners.OssIndexAnalysisTask] Starting Sonatype OSS Index analysis task
2022-06-28 07:23:21,187 [] INFO [org.dependencytrack.tasks.scanners.OssIndexAnalysisTask] Sonatype OSS Index analysis complete
2022-06-28 07:23:21,189 [] INFO [org.dependencytrack.policy.PolicyEngine] Evaluating 18477 component(s) against applicable policies
2022-06-28 07:24:34,036 [] INFO [org.dependencytrack.policy.PolicyEngine] Policy analysis complete
2022-06-28 07:24:34,038 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing metrics update for project: cd6feaf1-c6df-4c15-a7b4-f118151a6605
2022-06-28 07:24:41,373 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed metrics update for project: cd6feaf1-c6df-4c15-a7b4-f118151a6605
2022-06-28 07:25:37,850 [] INFO [org.dependencytrack.tasks.ClearComponentAnalysisCacheTask] Clearing ComponentAnalysisCache
2022-06-28 07:25:37,909 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing metrics update on vulnerability database
2022-06-28 07:25:37,933 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing portfolio metrics update
2022-06-28 07:25:37,945 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing metrics update for project: 7be7ded7-ac02-40c1-a3f8-ca11d1bc005f
2022-06-28 07:25:40,410 [] INFO [org.dependencytrack.tasks.ClearComponentAnalysisCacheTask] Complete
2022-06-28 07:25:46,910 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed metrics update for project: 7be7ded7-ac02-40c1-a3f8-ca11d1bc005f
2022-06-28 07:25:46,914 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing metrics update for project: cd6feaf1-c6df-4c15-a7b4-f118151a6605
2022-06-28 07:25:55,015 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed metrics update for project: cd6feaf1-c6df-4c15-a7b4-f118151a6605
2022-06-28 07:25:55,030 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed portfolio metrics update
2022-06-28 07:26:27,843 [] INFO [org.dependencytrack.tasks.VulnDbSyncTask] Starting VulnDB mirror synchronization task
2022-06-28 07:26:27,844 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Starting NIST mirroring task
2022-06-28 07:26:27,844 [] INFO [org.dependencytrack.tasks.VulnDbSyncTask] VulnDB mirror directory does not exist. Skipping.
2022-06-28 07:26:27,848 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Downloading files at Tue Jun 28 07:26:27 UTC 2022
2022-06-28 07:26:28,618 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Initiating download of https://nvd.nist.gov/feeds/json/cve/1.1/nvdcve-1.1-modified.json.gz
2022-06-28 07:26:28,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Downloading...
2022-06-28 07:26:29,602 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Uncompressing nvdcve-1.1-modified.json.gz
2022-06-28 07:26:29,711 [] INFO [org.dependencytrack.parser.nvd.NvdParser] Parsing nvdcve-1.1-modified.json
2022-06-28 07:30:44,409 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed metrics update on vulnerability database
2022-06-28 07:33:25,101 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Initiating download of https://nvd.nist.gov/feeds/json/cve/1.1/nvdcve-1.1-modified.meta
2022-06-28 07:33:25,964 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Downloading...
2022-06-28 07:33:25,972 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2022.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,972 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2021.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,972 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2020.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,972 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2019.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,972 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2018.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,973 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2017.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,973 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2016.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,973 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2015.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,973 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2014.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,973 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2013.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,973 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2012.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,973 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2011.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2010.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2009.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2008.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2007.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2006.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2005.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2004.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2003.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2002.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,975 [] INFO [org.dependencytrack.tasks.NistMirrorTask] NIST mirroring complete
2022-06-28 07:33:25,975 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Time spent (d/l): 1219ms
2022-06-28 07:33:25,975 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Time spent (parse): 415389ms
2022-06-28 07:33:25,975 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Time spent (total): 418131ms
2022-06-28 07:33:26,466 [] INFO [org.dependencytrack.tasks.EpssMirrorTask] Starting EPSS mirroring task
2022-06-28 07:33:26,466 [] INFO [org.dependencytrack.tasks.EpssMirrorTask] Initiating download of https://epss.cyentia.com/epss_scores-current.csv.gz
2022-06-28 07:33:27,918 [] INFO [org.dependencytrack.tasks.EpssMirrorTask] Downloading...
2022-06-28 07:33:28,448 [] INFO [org.dependencytrack.tasks.EpssMirrorTask] Uncompressing epss_scores-current.csv.gz
2022-06-28 07:33:28,533 [] INFO [org.dependencytrack.parser.epss.EpssParser] Parsing epss_scores-current.csv
2022-06-28 07:52:55,034 [] INFO [org.dependencytrack.tasks.EpssMirrorTask] EPSS mirroring complete
2022-06-28 07:52:55,034 [] INFO [org.dependencytrack.tasks.EpssMirrorTask] Time spent (d/l): 1452ms
2022-06-28 07:52:55,034 [] INFO [org.dependencytrack.tasks.EpssMirrorTask] Time spent (parse): 1166500ms
2022-06-28 07:52:55,034 [] INFO [org.dependencytrack.tasks.EpssMirrorTask] Time spent (total): 1168568ms
2022-06-28 08:25:27,852 [] INFO [org.dependencytrack.tasks.InternalComponentIdentificationTask] Starting internal component identification task
2022-06-28 08:25:27,914 [] INFO [org.dependencytrack.tasks.repositories.RepositoryMetaAnalyzerTask] Performing component repository metadata analysis against 564 components in project: 7be7ded7-ac02-40c1-a3f8-ca11d1bc005f
2022-06-28 08:25:29,630 [] INFO [org.dependencytrack.tasks.InternalComponentIdentificationTask] Internal component identification task completed
2022-06-28 08:25:37,909 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing metrics update on vulnerability database
2022-06-28 08:25:37,934 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing portfolio metrics update
2022-06-28 08:25:37,942 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing metrics update for project: 7be7ded7-ac02-40c1-a3f8-ca11d1bc005f
2022-06-28 08:25:48,722 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed metrics update for project: 7be7ded7-ac02-40c1-a3f8-ca11d1bc005f
2022-06-28 08:25:48,725 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing metrics update for project: cd6feaf1-c6df-4c15-a7b4-f118151a6605
2022-06-28 08:25:57,114 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed metrics update for project: cd6feaf1-c6df-4c15-a7b4-f118151a6605
2022-06-28 08:25:57,132 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed portfolio metrics update
2022-06-28 08:30:36,552 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed metrics update on vulnerability database
2022-06-28 08:34:26,719 [] INFO [org.dependencytrack.tasks.repositories.RepositoryMetaAnalyzerTask] Completed component repository metadata analysis against 564 components in project: 7be7ded7-ac02-40c1-a3f8-ca11d1bc005f
2022-06-28 08:34:26,757 [] INFO [org.dependencytrack.tasks.repositories.RepositoryMetaAnalyzerTask] Performing component repository metadata analysis against 564 components in project: cd6feaf1-c6df-4c15-a7b4-f118151a6605
2022-06-28 08:42:17,083 [] INFO [org.dependencytrack.tasks.repositories.RepositoryMetaAnalyzerTask] Completed component repository metadata analysis against 564 components in project: cd6feaf1-c6df-4c15-a7b4-f118151a6605
2022-06-28 08:42:17,086 [] INFO [org.dependencytrack.tasks.repositories.RepositoryMetaAnalyzerTask] Portfolio component repository metadata analysis complete
dtrack@159c5a33cb10:/data$
| 1.0 | Dependency track takes more than 30 minutes to analyze and Dependency tracker jenkins aborts due to polling time limti. - The defect may already be reported! Please search for the defect before creating one.
### Current Behavior:
Jenkins Dependecy tarck jobs fails since server takes more 30 minutes analyze the project.
### Steps to Reproduce:
Run jenkins Dependecy track job in Jenkins pipeline for java project which has more than 1000 components(These are java modules)
### Expected Behavior:
Jenkins PipleLine should not fail.
### Environment:
- Dependency-Track Version: 'latest'
- Distribution: [ Docker ]
- BOM Format & Version: "bomFormat": "CycloneDX",
"specVersion": "1.4",
"version": 1,
- Database Server: PostgreSQL
- Browser: Chrome
### Additional Details:
22-06-28 06:37:51,463 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed portfolio metrics update
2022-06-28 06:37:56,340 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed metrics update for project: cd6feaf1-c6df-4c15-a7b4-f118151a6605
2022-06-28 06:52:29,046 [] INFO [org.dependencytrack.tasks.BomUploadProcessingTask] Processing CycloneDX BOM uploaded to project: cd6feaf1-c6df-4c15-a7b4-f118151a6605
2022-06-28 06:56:59,202 [] INFO [org.dependencytrack.tasks.BomUploadProcessingTask] Processing CycloneDX dependency graph for project: cd6feaf1-c6df-4c15-a7b4-f118151a6605
2022-06-28 07:04:14,351 [] INFO [org.dependencytrack.tasks.BomUploadProcessingTask] Processed 18477 components and 0 services uploaded to project cd6feaf1-c6df-4c15-a7b4-f118151a6605
2022-06-28 07:22:08,107 [] INFO [org.dependencytrack.tasks.scanners.InternalAnalysisTask] Starting internal analysis task
2022-06-28 07:22:42,573 [] INFO [org.dependencytrack.tasks.scanners.InternalAnalysisTask] Internal analysis complete
2022-06-28 07:22:42,579 [] WARN [org.dependencytrack.tasks.scanners.OssIndexAnalysisTask] An API username or token has not been specified for use with OSS Index. Using anonymous access
2022-06-28 07:22:42,579 [] INFO [org.dependencytrack.tasks.scanners.OssIndexAnalysisTask] Starting Sonatype OSS Index analysis task
2022-06-28 07:23:21,187 [] INFO [org.dependencytrack.tasks.scanners.OssIndexAnalysisTask] Sonatype OSS Index analysis complete
2022-06-28 07:23:21,189 [] INFO [org.dependencytrack.policy.PolicyEngine] Evaluating 18477 component(s) against applicable policies
2022-06-28 07:24:34,036 [] INFO [org.dependencytrack.policy.PolicyEngine] Policy analysis complete
2022-06-28 07:24:34,038 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing metrics update for project: cd6feaf1-c6df-4c15-a7b4-f118151a6605
2022-06-28 07:24:41,373 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed metrics update for project: cd6feaf1-c6df-4c15-a7b4-f118151a6605
2022-06-28 07:25:37,850 [] INFO [org.dependencytrack.tasks.ClearComponentAnalysisCacheTask] Clearing ComponentAnalysisCache
2022-06-28 07:25:37,909 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing metrics update on vulnerability database
2022-06-28 07:25:37,933 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing portfolio metrics update
2022-06-28 07:25:37,945 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing metrics update for project: 7be7ded7-ac02-40c1-a3f8-ca11d1bc005f
2022-06-28 07:25:40,410 [] INFO [org.dependencytrack.tasks.ClearComponentAnalysisCacheTask] Complete
2022-06-28 07:25:46,910 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed metrics update for project: 7be7ded7-ac02-40c1-a3f8-ca11d1bc005f
2022-06-28 07:25:46,914 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing metrics update for project: cd6feaf1-c6df-4c15-a7b4-f118151a6605
2022-06-28 07:25:55,015 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed metrics update for project: cd6feaf1-c6df-4c15-a7b4-f118151a6605
2022-06-28 07:25:55,030 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed portfolio metrics update
2022-06-28 07:26:27,843 [] INFO [org.dependencytrack.tasks.VulnDbSyncTask] Starting VulnDB mirror synchronization task
2022-06-28 07:26:27,844 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Starting NIST mirroring task
2022-06-28 07:26:27,844 [] INFO [org.dependencytrack.tasks.VulnDbSyncTask] VulnDB mirror directory does not exist. Skipping.
2022-06-28 07:26:27,848 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Downloading files at Tue Jun 28 07:26:27 UTC 2022
2022-06-28 07:26:28,618 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Initiating download of https://nvd.nist.gov/feeds/json/cve/1.1/nvdcve-1.1-modified.json.gz
2022-06-28 07:26:28,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Downloading...
2022-06-28 07:26:29,602 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Uncompressing nvdcve-1.1-modified.json.gz
2022-06-28 07:26:29,711 [] INFO [org.dependencytrack.parser.nvd.NvdParser] Parsing nvdcve-1.1-modified.json
2022-06-28 07:30:44,409 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed metrics update on vulnerability database
2022-06-28 07:33:25,101 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Initiating download of https://nvd.nist.gov/feeds/json/cve/1.1/nvdcve-1.1-modified.meta
2022-06-28 07:33:25,964 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Downloading...
2022-06-28 07:33:25,972 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2022.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,972 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2021.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,972 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2020.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,972 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2019.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,972 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2018.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,973 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2017.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,973 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2016.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,973 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2015.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,973 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2014.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,973 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2013.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,973 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2012.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,973 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2011.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2010.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2009.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2008.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2007.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2006.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2005.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2004.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2003.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,974 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Retrieval of nvdcve-1.1-2002.json.gz not necessary. Will use modified feed for updates.
2022-06-28 07:33:25,975 [] INFO [org.dependencytrack.tasks.NistMirrorTask] NIST mirroring complete
2022-06-28 07:33:25,975 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Time spent (d/l): 1219ms
2022-06-28 07:33:25,975 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Time spent (parse): 415389ms
2022-06-28 07:33:25,975 [] INFO [org.dependencytrack.tasks.NistMirrorTask] Time spent (total): 418131ms
2022-06-28 07:33:26,466 [] INFO [org.dependencytrack.tasks.EpssMirrorTask] Starting EPSS mirroring task
2022-06-28 07:33:26,466 [] INFO [org.dependencytrack.tasks.EpssMirrorTask] Initiating download of https://epss.cyentia.com/epss_scores-current.csv.gz
2022-06-28 07:33:27,918 [] INFO [org.dependencytrack.tasks.EpssMirrorTask] Downloading...
2022-06-28 07:33:28,448 [] INFO [org.dependencytrack.tasks.EpssMirrorTask] Uncompressing epss_scores-current.csv.gz
2022-06-28 07:33:28,533 [] INFO [org.dependencytrack.parser.epss.EpssParser] Parsing epss_scores-current.csv
2022-06-28 07:52:55,034 [] INFO [org.dependencytrack.tasks.EpssMirrorTask] EPSS mirroring complete
2022-06-28 07:52:55,034 [] INFO [org.dependencytrack.tasks.EpssMirrorTask] Time spent (d/l): 1452ms
2022-06-28 07:52:55,034 [] INFO [org.dependencytrack.tasks.EpssMirrorTask] Time spent (parse): 1166500ms
2022-06-28 07:52:55,034 [] INFO [org.dependencytrack.tasks.EpssMirrorTask] Time spent (total): 1168568ms
2022-06-28 08:25:27,852 [] INFO [org.dependencytrack.tasks.InternalComponentIdentificationTask] Starting internal component identification task
2022-06-28 08:25:27,914 [] INFO [org.dependencytrack.tasks.repositories.RepositoryMetaAnalyzerTask] Performing component repository metadata analysis against 564 components in project: 7be7ded7-ac02-40c1-a3f8-ca11d1bc005f
2022-06-28 08:25:29,630 [] INFO [org.dependencytrack.tasks.InternalComponentIdentificationTask] Internal component identification task completed
2022-06-28 08:25:37,909 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing metrics update on vulnerability database
2022-06-28 08:25:37,934 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing portfolio metrics update
2022-06-28 08:25:37,942 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing metrics update for project: 7be7ded7-ac02-40c1-a3f8-ca11d1bc005f
2022-06-28 08:25:48,722 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed metrics update for project: 7be7ded7-ac02-40c1-a3f8-ca11d1bc005f
2022-06-28 08:25:48,725 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Executing metrics update for project: cd6feaf1-c6df-4c15-a7b4-f118151a6605
2022-06-28 08:25:57,114 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed metrics update for project: cd6feaf1-c6df-4c15-a7b4-f118151a6605
2022-06-28 08:25:57,132 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed portfolio metrics update
2022-06-28 08:30:36,552 [] INFO [org.dependencytrack.tasks.MetricsUpdateTask] Completed metrics update on vulnerability database
2022-06-28 08:34:26,719 [] INFO [org.dependencytrack.tasks.repositories.RepositoryMetaAnalyzerTask] Completed component repository metadata analysis against 564 components in project: 7be7ded7-ac02-40c1-a3f8-ca11d1bc005f
2022-06-28 08:34:26,757 [] INFO [org.dependencytrack.tasks.repositories.RepositoryMetaAnalyzerTask] Performing component repository metadata analysis against 564 components in project: cd6feaf1-c6df-4c15-a7b4-f118151a6605
2022-06-28 08:42:17,083 [] INFO [org.dependencytrack.tasks.repositories.RepositoryMetaAnalyzerTask] Completed component repository metadata analysis against 564 components in project: cd6feaf1-c6df-4c15-a7b4-f118151a6605
2022-06-28 08:42:17,086 [] INFO [org.dependencytrack.tasks.repositories.RepositoryMetaAnalyzerTask] Portfolio component repository metadata analysis complete
dtrack@159c5a33cb10:/data$
| defect | dependency track takes more than minutes to analyze and dependency tracker jenkins aborts due to polling time limti the defect may already be reported please search for the defect before creating one current behavior jenkins dependecy tarck jobs fails since server takes more minutes analyze the project steps to reproduce run jenkins dependecy track job in jenkins pipeline for java project which has more than components these are java modules expected behavior jenkins pipleline should not fail environment dependency track version latest distribution bom format version bomformat cyclonedx specversion version database server postgresql browser chrome additional details info completed portfolio metrics update info completed metrics update for project info processing cyclonedx bom uploaded to project info processing cyclonedx dependency graph for project info processed components and services uploaded to project info starting internal analysis task info internal analysis complete warn an api username or token has not been specified for use with oss index using anonymous access info starting sonatype oss index analysis task info sonatype oss index analysis complete info evaluating component s against applicable policies info policy analysis complete info executing metrics update for project info completed metrics update for project info clearing componentanalysiscache info executing metrics update on vulnerability database info executing portfolio metrics update info executing metrics update for project info complete info completed metrics update for project info executing metrics update for project info completed metrics update for project info completed portfolio metrics update info starting vulndb mirror synchronization task info starting nist mirroring task info vulndb mirror directory does not exist skipping info downloading files at tue jun utc info initiating download of info downloading info uncompressing nvdcve modified json gz info parsing nvdcve modified json info completed metrics update on vulnerability database info initiating download of info downloading info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info retrieval of nvdcve json gz not necessary will use modified feed for updates info nist mirroring complete info time spent d l info time spent parse info time spent total info starting epss mirroring task info initiating download of info downloading info uncompressing epss scores current csv gz info parsing epss scores current csv info epss mirroring complete info time spent d l info time spent parse info time spent total info starting internal component identification task info performing component repository metadata analysis against components in project info internal component identification task completed info executing metrics update on vulnerability database info executing portfolio metrics update info executing metrics update for project info completed metrics update for project info executing metrics update for project info completed metrics update for project info completed portfolio metrics update info completed metrics update on vulnerability database info completed component repository metadata analysis against components in project info performing component repository metadata analysis against components in project info completed component repository metadata analysis against components in project info portfolio component repository metadata analysis complete dtrack data | 1 |
566,673 | 16,826,842,335 | IssuesEvent | 2021-06-17 19:50:22 | rokwire/illinois-app | https://api.github.com/repos/rokwire/illinois-app | opened | [BUG] Career center link is not working | Priority: High Type: Bug | **Describe the bug**
Student Guide - The Career center link is not doing anything when the user taps on it.
**To Reproduce**
Steps to reproduce the behavior:
1. Install the Illinois app
2. Complete the onboarding and verify the role as a student role
3. Via browse screen, navigate to Student guide and navigate to Career -> Get Ready for the job Market
4. Tap on Career Center.
**Actual Screen**
The Career center link is not doing anything when the user taps on it.
**Expected behavior**
The Career center webpage should load when the user tap on the link*
**Screenshots**


**Smartphone (please complete the following information):**
- Device: [e.g. Android]
- Version [e.g. 2.4.25]
| 1.0 | [BUG] Career center link is not working - **Describe the bug**
Student Guide - The Career center link is not doing anything when the user taps on it.
**To Reproduce**
Steps to reproduce the behavior:
1. Install the Illinois app
2. Complete the onboarding and verify the role as a student role
3. Via browse screen, navigate to Student guide and navigate to Career -> Get Ready for the job Market
4. Tap on Career Center.
**Actual Screen**
The Career center link is not doing anything when the user taps on it.
**Expected behavior**
The Career center webpage should load when the user tap on the link*
**Screenshots**


**Smartphone (please complete the following information):**
- Device: [e.g. Android]
- Version [e.g. 2.4.25]
| non_defect | career center link is not working describe the bug student guide the career center link is not doing anything when the user taps on it to reproduce steps to reproduce the behavior install the illinois app complete the onboarding and verify the role as a student role via browse screen navigate to student guide and navigate to career get ready for the job market tap on career center actual screen the career center link is not doing anything when the user taps on it expected behavior the career center webpage should load when the user tap on the link screenshots smartphone please complete the following information device version | 0 |
5,674 | 2,610,193,081 | IssuesEvent | 2015-02-26 19:00:58 | chrsmith/quchuseban | https://api.github.com/repos/chrsmith/quchuseban | opened | 推荐脸上长色斑去除偏方 | auto-migrated Priority-Medium Type-Defect | ```
《摘要》
四季的轮回,变换着多样缤纷的色彩。当黄,静静的铺满山坡。一丝怅然恐慌成无法言说的失落。凝不住经年的妩媚,留不下往昔的绚烂。在季节的深处,九月的风微凉着心事。一场寒雨在风中呜咽叹息,轻敲轩窗,碎断人肠。没有了昨日纤雨湿花的雅致,忘却了曾经柳絮扑帘的悠然。绮梦悠悠伤情绵绵,一任愁绪在风中雨里飘摇成伤,恍若,枝头颤动的叶子,没了昨日张扬的碧绿,沉淀了一个春夏的水嫩在风中瞬间走失,只留下一袭苍白,挣扎在季节的末梢。嘴角溢出的笑,潋滟了一湖的秋水。如果出去嘴角的色斑,我想这个微笑会更美.脸上长色斑去除偏方,
《客户案例》
莫小姐 24岁<br>
长时间面对电脑的坏处我一直都知道,像眼睛疲劳、干��
�,身体处于亚健康之类的,工作一年来我基本上10个小时对��
�电脑,好在我什么不舒服的症状都没出现,正当我庆幸不已�
��同时,我发现我额头、鼻梁两边和上眼睑处长斑了,数量很
多,就像雨后春笋一样,一点一点在短时间内全部冒出来了��
�看着脸上那如同苍蝇屎大小的斑点,颜色为黄色,略比肤色�
��点,远看近看都明显。哪个女孩子不爱美?为此特别烦恼。<b
r>
在网上搜索祛斑产品时发现了「黛芙薇尔精华液」,了��
�了一番之后觉得还不错,而且网上论坛里面大家发帖子都说�
��黛芙薇尔精华液」好,没发现有说副作用的,然后我还上「
黛芙薇尔精华液」的网站很认真的看了一下产品说明,说明��
�字很专业详细,对治疗机理、产品优势描述得挺不错的,我�
��在网上订购了一个周期的。<br>
大概了十几天的样子,我鼻梁两边的斑块颜色就变淡了��
�面积也慢慢在缩小,本来粗糙、多油的皮肤状况也得到了改�
��。一个月过去以后,对着镜子看自己脸上的斑,鼻梁上的大
斑块已经缩小了一大半,颜色由以前的深褐色变成了浅褐色��
�额头和眼角处的斑也变淡变浅了。「黛芙薇尔精华液」祛斑�
��家在我订购产品的时候就说过,产品在祛斑的同时还能调经
,让经期变得正常,还能减轻痛经的痛苦。我使用第二个周��
�产品的时候,这方面的效果显示出来了,我以前痛经非常厉�
��,自从使用这个套装以来,不但月经量变得正常了,而且痛
经也没那么严重了,在祛斑的同时还能调经,「黛芙薇尔精��
�液」真的很不错。
很快我就使用完第二个周期的产品,鼻梁两边的斑基本上看��
�到了,颜色跟肤色很接近,不仔细看根本看不出来了,我可�
��不用粉底盖住那一块的斑。额头上的斑也在慢慢消退,特别
是眼睑处的斑,淡化的比较明显。我接着使用第三个周期的��
�品,这个周期见效很快,刚半个月多一点,眼睑处的斑就消�
��了,褐色的斑融进皮肤里面,一点都看不出来了,额头上的
也差不多没有了。使用完三个周期的「黛芙薇尔精华液」之��
�,我从斑女人变成了一个水嫩白净的女人,而且我的月经也�
��得很有规律,内分泌正常,心情也变得非常开朗,哪位想要
祛斑的朋友就学我的祛斑方法,相信通过使用「黛芙薇尔精��
�液」,大家都能祛斑成功。
阅读了脸上长色斑去除偏方,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
脸上长色斑去除偏方,同时为您分享祛斑小方法
一本中医古书上看到了一则去除雀斑的方子:将黄豆(黄豆��
�生的)浸泡在醋中一个月,每天服用几粒,坚持一段时间即�
��完全消除。 (去斑讲究内外兼修,这点值得一试) 。
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 5:38 | 1.0 | 推荐脸上长色斑去除偏方 - ```
《摘要》
四季的轮回,变换着多样缤纷的色彩。当黄,静静的铺满山坡。一丝怅然恐慌成无法言说的失落。凝不住经年的妩媚,留不下往昔的绚烂。在季节的深处,九月的风微凉着心事。一场寒雨在风中呜咽叹息,轻敲轩窗,碎断人肠。没有了昨日纤雨湿花的雅致,忘却了曾经柳絮扑帘的悠然。绮梦悠悠伤情绵绵,一任愁绪在风中雨里飘摇成伤,恍若,枝头颤动的叶子,没了昨日张扬的碧绿,沉淀了一个春夏的水嫩在风中瞬间走失,只留下一袭苍白,挣扎在季节的末梢。嘴角溢出的笑,潋滟了一湖的秋水。如果出去嘴角的色斑,我想这个微笑会更美.脸上长色斑去除偏方,
《客户案例》
莫小姐 24岁<br>
长时间面对电脑的坏处我一直都知道,像眼睛疲劳、干��
�,身体处于亚健康之类的,工作一年来我基本上10个小时对��
�电脑,好在我什么不舒服的症状都没出现,正当我庆幸不已�
��同时,我发现我额头、鼻梁两边和上眼睑处长斑了,数量很
多,就像雨后春笋一样,一点一点在短时间内全部冒出来了��
�看着脸上那如同苍蝇屎大小的斑点,颜色为黄色,略比肤色�
��点,远看近看都明显。哪个女孩子不爱美?为此特别烦恼。<b
r>
在网上搜索祛斑产品时发现了「黛芙薇尔精华液」,了��
�了一番之后觉得还不错,而且网上论坛里面大家发帖子都说�
��黛芙薇尔精华液」好,没发现有说副作用的,然后我还上「
黛芙薇尔精华液」的网站很认真的看了一下产品说明,说明��
�字很专业详细,对治疗机理、产品优势描述得挺不错的,我�
��在网上订购了一个周期的。<br>
大概了十几天的样子,我鼻梁两边的斑块颜色就变淡了��
�面积也慢慢在缩小,本来粗糙、多油的皮肤状况也得到了改�
��。一个月过去以后,对着镜子看自己脸上的斑,鼻梁上的大
斑块已经缩小了一大半,颜色由以前的深褐色变成了浅褐色��
�额头和眼角处的斑也变淡变浅了。「黛芙薇尔精华液」祛斑�
��家在我订购产品的时候就说过,产品在祛斑的同时还能调经
,让经期变得正常,还能减轻痛经的痛苦。我使用第二个周��
�产品的时候,这方面的效果显示出来了,我以前痛经非常厉�
��,自从使用这个套装以来,不但月经量变得正常了,而且痛
经也没那么严重了,在祛斑的同时还能调经,「黛芙薇尔精��
�液」真的很不错。
很快我就使用完第二个周期的产品,鼻梁两边的斑基本上看��
�到了,颜色跟肤色很接近,不仔细看根本看不出来了,我可�
��不用粉底盖住那一块的斑。额头上的斑也在慢慢消退,特别
是眼睑处的斑,淡化的比较明显。我接着使用第三个周期的��
�品,这个周期见效很快,刚半个月多一点,眼睑处的斑就消�
��了,褐色的斑融进皮肤里面,一点都看不出来了,额头上的
也差不多没有了。使用完三个周期的「黛芙薇尔精华液」之��
�,我从斑女人变成了一个水嫩白净的女人,而且我的月经也�
��得很有规律,内分泌正常,心情也变得非常开朗,哪位想要
祛斑的朋友就学我的祛斑方法,相信通过使用「黛芙薇尔精��
�液」,大家都能祛斑成功。
阅读了脸上长色斑去除偏方,再看脸上容易长斑的原因:
《色斑形成原因》
内部因素
一、压力
当人受到压力时,就会分泌肾上腺素,为对付压力而做��
�备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏�
��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃
。
二、荷尔蒙分泌失调
避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞��
�分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在�
��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕
中因女性荷尔蒙雌激素的增加,从怀孕4—5个月开始会容易出
现斑,这时候出现的斑点在产后大部分会消失。可是,新陈��
�谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等�
��因,都会使斑加深。有时新长出的斑,产后也不会消失,所
以需要更加注意。
三、新陈代谢缓慢
肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑��
�因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态�
��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是
内分泌失调导致过敏体质而形成的。另外,身体状态不正常��
�时候,紫外线的照射也会加速斑的形成。
四、错误的使用化妆品
使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在��
�疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵�
��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的
问题。
外部因素
一、紫外线
照射紫外线的时候,人体为了保护皮肤,会在基底层产��
�很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更�
��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化,
还会引起黑斑、雀斑等色素沉着的皮肤疾患。
二、不良的清洁习惯
因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。��
�皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦�
��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的
问题。
三、遗传基因
父母中有长斑的,则本人长斑的概率就很高,这种情况��
�一定程度上就可判定是遗传基因的作用。所以家里特别是长�
��有长斑的人,要注意避免引发长斑的重要因素之一——紫外
线照射,这是预防斑必须注意的。
《有疑问帮你解决》
1,黛芙薇尔精华液真的有效果吗?真的可以把脸上的黄褐��
�去掉吗?
答:黛芙薇尔精华液DNA精华能够有效的修复周围难以触��
�的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必�
��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑
,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时��
�,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的�
��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显
而易见。自产品上市以来,老顾客纷纷介绍新顾客,71%的新��
�客都是通过老顾客介绍而来,口碑由此而来!
2,服用黛芙薇尔美白,会伤身体吗?有副作用吗?
答:黛芙薇尔精华液应用了精纯复合配方和领先的分类��
�斑科技,并将“DNA美肤系统”疗法应用到了该产品中,能彻�
��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有
效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾��
�地的专家通力协作,超过10年的研究以全新的DNA肌肤修复技��
�,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽�
��迹,令每一位爱美的女性都能享受到科技创新所带来的自然
之美。
专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数��
�百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖!
3,去除黄褐斑之后,会反弹吗?
答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔��
�白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家�
��据斑的形成原因精心研制而成用事实说话,让消费者打分。
树立权威品牌!我们的很多新客户都是老客户介绍而来,请问�
��如果效果不好,会有客户转介绍吗?
4,你们的价格有点贵,能不能便宜一点?
答:如果您使用西药最少需要2000元,煎服的药最少需要3
000元,做手术最少是5000元,而这些毫无疑问,不会对彻底去�
��你的斑点有任何帮助!一分价钱,一份价值,我们现在做的��
�是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的�
��褐斑彻底去除,你还会觉得贵吗?你还会再去花那么多冤枉��
�,不但斑没去掉,还把自己的皮肤弄的越来越糟吗
5,我适合用黛芙薇尔精华液吗?
答:黛芙薇尔适用人群:
1、生理紊乱引起的黄褐斑人群
2、生育引起的妊娠斑人群
3、年纪增长引起的老年斑人群
4、化妆品色素沉积、辐射斑人群
5、长期日照引起的日晒斑人群
6、肌肤暗淡急需美白的人群
《祛斑小方法》
脸上长色斑去除偏方,同时为您分享祛斑小方法
一本中医古书上看到了一则去除雀斑的方子:将黄豆(黄豆��
�生的)浸泡在醋中一个月,每天服用几粒,坚持一段时间即�
��完全消除。 (去斑讲究内外兼修,这点值得一试) 。
```
-----
Original issue reported on code.google.com by `additive...@gmail.com` on 1 Jul 2014 at 5:38 | defect | 推荐脸上长色斑去除偏方 《摘要》 四季的轮回,变换着多样缤纷的色彩。当黄,静静的铺满山坡。一丝怅然恐慌成无法言说的失落。凝不住经年的妩媚,留不下往昔的绚烂。在季节的深处,九月的风微凉着心事。一场寒雨在风中呜咽叹息,轻敲轩窗,碎断人肠。没有了昨日纤雨湿花的雅致,忘却了曾经柳絮扑帘的悠然。绮梦悠悠伤情绵绵,一任愁绪在风中雨里飘摇成伤,恍若,枝头颤动的叶子,没了昨日张扬的碧绿,沉淀了一个春夏的水嫩在风中瞬间走失,只留下一袭苍白,挣扎在季节的末梢。嘴角溢出的笑,潋滟了一湖的秋水。如果出去嘴角的色斑,我想这个微笑会更美 脸上长色斑去除偏方, 《客户案例》 莫小姐 长时间面对电脑的坏处我一直都知道,像眼睛疲劳、干�� �,身体处于亚健康之类的, �� �电脑,好在我什么不舒服的症状都没出现,正当我庆幸不已� ��同时,我发现我额头、鼻梁两边和上眼睑处长斑了,数量很 多,就像雨后春笋一样,一点一点在短时间内全部冒出来了�� �看着脸上那如同苍蝇屎大小的斑点,颜色为黄色,略比肤色� ��点,远看近看都明显。哪个女孩子不爱美 为此特别烦恼。 b r 在网上搜索祛斑产品时发现了「黛芙薇尔精华液」,了�� �了一番之后觉得还不错,而且网上论坛里面大家发帖子都说� ��黛芙薇尔精华液」好,没发现有说副作用的,然后我还上「 黛芙薇尔精华液」的网站很认真的看了一下产品说明,说明�� �字很专业详细,对治疗机理、产品优势描述得挺不错的,我� ��在网上订购了一个周期的。 大概了十几天的样子,我鼻梁两边的斑块颜色就变淡了�� �面积也慢慢在缩小,本来粗糙、多油的皮肤状况也得到了改� ��。一个月过去以后,对着镜子看自己脸上的斑,鼻梁上的大 斑块已经缩小了一大半,颜色由以前的深褐色变成了浅褐色�� �额头和眼角处的斑也变淡变浅了。「黛芙薇尔精华液」祛斑� ��家在我订购产品的时候就说过,产品在祛斑的同时还能调经 ,让经期变得正常,还能减轻痛经的痛苦。我使用第二个周�� �产品的时候,这方面的效果显示出来了,我以前痛经非常厉� ��,自从使用这个套装以来,不但月经量变得正常了,而且痛 经也没那么严重了,在祛斑的同时还能调经,「黛芙薇尔精�� �液」真的很不错。 很快我就使用完第二个周期的产品,鼻梁两边的斑基本上看�� �到了,颜色跟肤色很接近,不仔细看根本看不出来了,我可� ��不用粉底盖住那一块的斑。额头上的斑也在慢慢消退,特别 是眼睑处的斑,淡化的比较明显。我接着使用第三个周期的�� �品,这个周期见效很快,刚半个月多一点,眼睑处的斑就消� ��了,褐色的斑融进皮肤里面,一点都看不出来了,额头上的 也差不多没有了。使用完三个周期的「黛芙薇尔精华液」之�� �,我从斑女人变成了一个水嫩白净的女人,而且我的月经也� ��得很有规律,内分泌正常,心情也变得非常开朗,哪位想要 祛斑的朋友就学我的祛斑方法,相信通过使用「黛芙薇尔精�� �液」,大家都能祛斑成功。 阅读了脸上长色斑去除偏方,再看脸上容易长斑的原因: 《色斑形成原因》 内部因素 一、压力 当人受到压力时,就会分泌肾上腺素,为对付压力而做�� �备。如果长期受到压力,人体新陈代谢的平衡就会遭到破坏� ��皮肤所需的营养供应趋于缓慢,色素母细胞就会变得很活跃 。 二、荷尔蒙分泌失调 避孕药里所含的女性荷尔蒙雌激素,会刺激麦拉宁细胞�� �分泌而形成不均匀的斑点,因避孕药而形成的斑点,虽然在� ��药中断后会停止,但仍会在皮肤上停留很长一段时间。怀孕 中因女性荷尔蒙雌激素的增加, — 现斑,这时候出现的斑点在产后大部分会消失。可是,新陈�� �谢不正常、肌肤裸露在强烈的紫外线下、精神上受到压力等� ��因,都会使斑加深。有时新长出的斑,产后也不会消失,所 以需要更加注意。 三、新陈代谢缓慢 肝的新陈代谢功能不正常或卵巢功能减退时也会出现斑�� �因为新陈代谢不顺畅、或内分泌失调,使身体处于敏感状态� ��,从而加剧色素问题。我们常说的便秘会形成斑,其实就是 内分泌失调导致过敏体质而形成的。另外,身体状态不正常�� �时候,紫外线的照射也会加速斑的形成。 四、错误的使用化妆品 使用了不适合自己皮肤的化妆品,会导致皮肤过敏。在�� �疗的过程中如过量照射到紫外线,皮肤会为了抵御外界的侵� ��,在有炎症的部位聚集麦拉宁色素,这样会出现色素沉着的 问题。 外部因素 一、紫外线 照射紫外线的时候,人体为了保护皮肤,会在基底层产�� �很多麦拉宁色素。所以为了保护皮肤,会在敏感部位聚集更� ��的色素。经常裸露在强烈的阳光底下不仅促进皮肤的老化, 还会引起黑斑、雀斑等色素沉着的皮肤疾患。 二、不良的清洁习惯 因强烈的清洁习惯使皮肤变得敏感,这样会刺激皮肤。�� �皮肤敏感时,人体为了保护皮肤,黑色素细胞会分泌很多麦� ��宁色素,当色素过剩时就出现了斑、瑕疵等皮肤色素沉着的 问题。 三、遗传基因 父母中有长斑的,则本人长斑的概率就很高,这种情况�� �一定程度上就可判定是遗传基因的作用。所以家里特别是长� ��有长斑的人,要注意避免引发长斑的重要因素之一——紫外 线照射,这是预防斑必须注意的。 《有疑问帮你解决》 黛芙薇尔精华液真的有效果吗 真的可以把脸上的黄褐�� �去掉吗 答:黛芙薇尔精华液dna精华能够有效的修复周围难以触�� �的色斑,其独有的纳豆成分为皮肤的美白与靓丽,提供了必� ��可少的营养物质,可以有效的去除黄褐斑,黄褐斑,黄褐斑 ,蝴蝶斑,晒斑、妊娠斑等。它它完全突破了传统的美肤时�� �,宛如在皮肤中注入了一杯兼具活化、再生、滋养等功效的� ��尾酒,同时为脸部提供大量有机维生素精华,脸部的改变显 而易见。自产品上市以来,老顾客纷纷介绍新顾客, 的新�� �客都是通过老顾客介绍而来,口碑由此而来 ,服用黛芙薇尔美白,会伤身体吗 有副作用吗 答:黛芙薇尔精华液应用了精纯复合配方和领先的分类�� �斑科技,并将“dna美肤系统”疗法应用到了该产品中,能彻� ��祛除黄褐斑,蝴蝶斑,妊娠斑,晒斑,黄褐斑,老年斑,有 效淡化黄褐斑至接近肤色。黛芙薇尔通过法国、美国、台湾�� �地的专家通力协作, �� �,挑战传统化学护肤理念,不懈追寻发现破译大自然的美丽� ��迹,令每一位爱美的女性都能享受到科技创新所带来的自然 之美。 专为亚洲女性肤质研制,精心呵护女性美丽,多年来,为数�� �百万计的女性解除了黄褐斑困扰。深得广大女性朋友的信赖 ,去除黄褐斑之后,会反弹吗 答:很多曾经长了黄褐斑的人士,自从选择了黛芙薇尔�� �白,就一劳永逸。这款祛斑产品是经过数十位权威祛斑专家� ��据斑的形成原因精心研制而成用事实说话,让消费者打分。 树立权威品牌 我们的很多新客户都是老客户介绍而来,请问� ��如果效果不好,会有客户转介绍吗 ,你们的价格有点贵,能不能便宜一点 答: , , ,而这些毫无疑问,不会对彻底去� ��你的斑点有任何帮助 一分价钱,一份价值,我们现在做的�� �是一个口碑,一个品牌,价钱并不高。如果花这点钱把你的� ��褐斑彻底去除,你还会觉得贵吗 你还会再去花那么多冤枉�� �,不但斑没去掉,还把自己的皮肤弄的越来越糟吗 ,我适合用黛芙薇尔精华液吗 答:黛芙薇尔适用人群: 、生理紊乱引起的黄褐斑人群 、生育引起的妊娠斑人群 、年纪增长引起的老年斑人群 、化妆品色素沉积、辐射斑人群 、长期日照引起的日晒斑人群 、肌肤暗淡急需美白的人群 《祛斑小方法》 脸上长色斑去除偏方,同时为您分享祛斑小方法 一本中医古书上看到了一则去除雀斑的方子:将黄豆(黄豆�� �生的)浸泡在醋中一个月,每天服用几粒,坚持一段时间即� ��完全消除。 去斑讲究内外兼修,这点值得一试 。 original issue reported on code google com by additive gmail com on jul at | 1 |
29,212 | 5,611,884,353 | IssuesEvent | 2017-04-03 01:21:31 | prettydiff/prettydiff | https://api.github.com/repos/prettydiff/prettydiff | opened | New simplified diff algorithm broke diffcli output | Defect Diff Not started | Currently the diffcli option is over-engineered to compensate for a poor diff algorithm. The formatting and building of output needs to be an array of output prepared in the diffview.js library so that consuming interfaces simply loop through this array output with a console.log statement and without any additional logic. | 1.0 | New simplified diff algorithm broke diffcli output - Currently the diffcli option is over-engineered to compensate for a poor diff algorithm. The formatting and building of output needs to be an array of output prepared in the diffview.js library so that consuming interfaces simply loop through this array output with a console.log statement and without any additional logic. | defect | new simplified diff algorithm broke diffcli output currently the diffcli option is over engineered to compensate for a poor diff algorithm the formatting and building of output needs to be an array of output prepared in the diffview js library so that consuming interfaces simply loop through this array output with a console log statement and without any additional logic | 1 |
29,928 | 5,959,397,176 | IssuesEvent | 2017-05-29 10:55:26 | bridgedotnet/Bridge | https://api.github.com/repos/bridgedotnet/Bridge | closed | Enum.TryParse fails with exception | defect in progress | A description of the issue.
### Steps To Reproduce
https://deck.net/08c324d3c077c47c7dade0e01405f1ec
https://dotnetfiddle.net/Runi2c
```c#
public enum Mode
{
None,
A,
B,
C
}
public class Program
{
private static void Main(string[] args)
{
const string section = null;
Mode mode;
if (Enum.TryParse(section, true, out mode) == false)
mode = Mode.A;
Console.WriteLine(mode);
}
}
```
### Expected Result
```js
A
```
### Actual Result
```js
TypeError: Cannot read property 'toLowerCase' of null
```
| 1.0 | Enum.TryParse fails with exception - A description of the issue.
### Steps To Reproduce
https://deck.net/08c324d3c077c47c7dade0e01405f1ec
https://dotnetfiddle.net/Runi2c
```c#
public enum Mode
{
None,
A,
B,
C
}
public class Program
{
private static void Main(string[] args)
{
const string section = null;
Mode mode;
if (Enum.TryParse(section, true, out mode) == false)
mode = Mode.A;
Console.WriteLine(mode);
}
}
```
### Expected Result
```js
A
```
### Actual Result
```js
TypeError: Cannot read property 'toLowerCase' of null
```
| defect | enum tryparse fails with exception a description of the issue steps to reproduce c public enum mode none a b c public class program private static void main string args const string section null mode mode if enum tryparse section true out mode false mode mode a console writeline mode expected result js a actual result js typeerror cannot read property tolowercase of null | 1 |
35,606 | 7,787,962,456 | IssuesEvent | 2018-06-07 01:35:55 | google/sanitizers | https://api.github.com/repos/google/sanitizers | closed | Fine-grained origins for class members | Priority-Medium ProjectMemorySanitizer Status-Accepted Type-Defect | Originally reported on Google Code with ID 31
```
Currently MSan reports the same origin for all members of dynamically allocated objects
(pointing to the new() call for that object).
It would be nice to include the actual member name in MSan report.
Looks like we need Clang support for this - llvm IR does not have enough info (there
are class layouts, but no member names).
```
Reported by `eugenis@google.com` on 2013-09-24 09:13:26
| 1.0 | Fine-grained origins for class members - Originally reported on Google Code with ID 31
```
Currently MSan reports the same origin for all members of dynamically allocated objects
(pointing to the new() call for that object).
It would be nice to include the actual member name in MSan report.
Looks like we need Clang support for this - llvm IR does not have enough info (there
are class layouts, but no member names).
```
Reported by `eugenis@google.com` on 2013-09-24 09:13:26
| defect | fine grained origins for class members originally reported on google code with id currently msan reports the same origin for all members of dynamically allocated objects pointing to the new call for that object it would be nice to include the actual member name in msan report looks like we need clang support for this llvm ir does not have enough info there are class layouts but no member names reported by eugenis google com on | 1 |
10,458 | 2,622,161,945 | IssuesEvent | 2015-03-04 00:10:25 | byzhang/terrastore | https://api.github.com/repos/byzhang/terrastore | closed | Backup implementation for Ensembles | auto-migrated Milestone-0.8.2 Priority-High Project-Terrastore Type-Defect | ```
How do you scale a currently running ensemble from N clusters to N+1
cluster without losing data?
I had a ensemble running of 2 clusters (each with 1 master and 1
server)
I tried the following, but could not retain all the data (I had just 1
bucket):
1. Export backup of the bucket of both the servers.
2. Restart the servers using the update cluster.json files. So now I
have a ensemble of 3 cluster (the third cluster is also 1 master and 1
server)
3. Import the backup on both the servers.
4. When I try to GET all data, it shows 404 for some of the records.
I even tried cleaning the masters before joining the servers in the
new ensemble, but faced the same issue.
```
Original issue reported on code.google.com by `rohi...@gmail.com` on 10 Jun 2011 at 2:58 | 1.0 | Backup implementation for Ensembles - ```
How do you scale a currently running ensemble from N clusters to N+1
cluster without losing data?
I had a ensemble running of 2 clusters (each with 1 master and 1
server)
I tried the following, but could not retain all the data (I had just 1
bucket):
1. Export backup of the bucket of both the servers.
2. Restart the servers using the update cluster.json files. So now I
have a ensemble of 3 cluster (the third cluster is also 1 master and 1
server)
3. Import the backup on both the servers.
4. When I try to GET all data, it shows 404 for some of the records.
I even tried cleaning the masters before joining the servers in the
new ensemble, but faced the same issue.
```
Original issue reported on code.google.com by `rohi...@gmail.com` on 10 Jun 2011 at 2:58 | defect | backup implementation for ensembles how do you scale a currently running ensemble from n clusters to n cluster without losing data i had a ensemble running of clusters each with master and server i tried the following but could not retain all the data i had just bucket export backup of the bucket of both the servers restart the servers using the update cluster json files so now i have a ensemble of cluster the third cluster is also master and server import the backup on both the servers when i try to get all data it shows for some of the records i even tried cleaning the masters before joining the servers in the new ensemble but faced the same issue original issue reported on code google com by rohi gmail com on jun at | 1 |
135,454 | 12,684,739,392 | IssuesEvent | 2020-06-19 23:57:58 | Azure/azure-cli | https://api.github.com/repos/Azure/azure-cli | closed | It's always good to add actual example of each command. | Documentation OKR3.4 Candidate | It's always good to add actual example of each command.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 9abfaa54-0895-29f7-f3bb-ff604912b2ac
* Version Independent ID: d1ff3478-8a86-59e4-62ef-449900132a67
* Content: [az account](https://docs.microsoft.com/en-us/cli/azure/account?view=azure-cli-latest#az-account-set)
* Content Source: [src/azure-cli/azure/cli/command_modules/profile/_help.py](https://github.com/Azure/azure-cli/blob/dev/src/azure-cli/azure/cli/command_modules/profile/_help.py)
* Service: **azure**
* GitHub Login: @rloutlaw
* Microsoft Alias: **routlaw** | 1.0 | It's always good to add actual example of each command. - It's always good to add actual example of each command.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 9abfaa54-0895-29f7-f3bb-ff604912b2ac
* Version Independent ID: d1ff3478-8a86-59e4-62ef-449900132a67
* Content: [az account](https://docs.microsoft.com/en-us/cli/azure/account?view=azure-cli-latest#az-account-set)
* Content Source: [src/azure-cli/azure/cli/command_modules/profile/_help.py](https://github.com/Azure/azure-cli/blob/dev/src/azure-cli/azure/cli/command_modules/profile/_help.py)
* Service: **azure**
* GitHub Login: @rloutlaw
* Microsoft Alias: **routlaw** | non_defect | it s always good to add actual example of each command it s always good to add actual example of each command document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service azure github login rloutlaw microsoft alias routlaw | 0 |
10,513 | 2,622,169,415 | IssuesEvent | 2015-03-04 00:13:49 | byzhang/rapidjson | https://api.github.com/repos/byzhang/rapidjson | closed | Unable to build tests on Mac OS X | auto-migrated Priority-Medium Type-Defect | ```
I am trying to build the tests for rapidjson 0.11
(http://code.google.com/p/rapidjson/) on Mac OS X . It includes three projects:
gtest (builds fine), unittest (build fails), and perftest (build fails), and
when building make error out with Error 1 and Error 2. Actually, the first time
I ran make, gtest and unittest built fine, but perftest failed. Now, unittest
is failing as well.
The compiler output shows the following errors for both unittest and perftest
which causes make to fail:
../../include/rapidjson/reader.h: In function ‘const char*
rapidjson::SkipWhitespace_SIMD(const char*)’:
../../include/rapidjson/reader.h:116: error: ‘_SIDD_UBYTE_OPS’ was not
declared in this scope
../../include/rapidjson/reader.h:116: error: ‘_SIDD_CMP_EQUAL_ANY’ was not
declared in this scope
../../include/rapidjson/reader.h:116: error: ‘_SIDD_BIT_MASK’ was not
declared in this scope
../../include/rapidjson/reader.h:116: error: ‘_SIDD_NEGATIVE_POLARITY’ was
not declared in this scope
These pre-processor constants are related to SSE4 instructions. The makefile
includes the -msse4.2 compiler switch to enable SSE4.2 support, and looking
through the header files reveal that on OS X, both SSE4_1 and SSE4_2
pre-processor constants need to be defined for the SIDD... constants to be
defined. For some reason, these SIDD... constants aren't being defined.
Further research showed that the -msse4 switch enables support for both SSE4.1
and SSE4.2, so I tried changing the switch to -msse4, but it still errors out.
Not sure if the -msse4.2 switch automatically defines SSE4_2__ , but I tried
manually defining it, and sill no luck.
NOTE: If you want to try building it yourself on Mac, you will need to download
a different premake script file, as the included one doesn't work. You can
download the corrected script from the attachment on the second post here
https://code.google.com/p/rapidjson/issues/detail?id=54
Any ideas on how to get it building successfully on OS X ?
```
Original issue reported on code.google.com by `keith3...@rocketmail.com` on 23 Jan 2013 at 5:28
* Merged into: #55 | 1.0 | Unable to build tests on Mac OS X - ```
I am trying to build the tests for rapidjson 0.11
(http://code.google.com/p/rapidjson/) on Mac OS X . It includes three projects:
gtest (builds fine), unittest (build fails), and perftest (build fails), and
when building make error out with Error 1 and Error 2. Actually, the first time
I ran make, gtest and unittest built fine, but perftest failed. Now, unittest
is failing as well.
The compiler output shows the following errors for both unittest and perftest
which causes make to fail:
../../include/rapidjson/reader.h: In function ‘const char*
rapidjson::SkipWhitespace_SIMD(const char*)’:
../../include/rapidjson/reader.h:116: error: ‘_SIDD_UBYTE_OPS’ was not
declared in this scope
../../include/rapidjson/reader.h:116: error: ‘_SIDD_CMP_EQUAL_ANY’ was not
declared in this scope
../../include/rapidjson/reader.h:116: error: ‘_SIDD_BIT_MASK’ was not
declared in this scope
../../include/rapidjson/reader.h:116: error: ‘_SIDD_NEGATIVE_POLARITY’ was
not declared in this scope
These pre-processor constants are related to SSE4 instructions. The makefile
includes the -msse4.2 compiler switch to enable SSE4.2 support, and looking
through the header files reveal that on OS X, both SSE4_1 and SSE4_2
pre-processor constants need to be defined for the SIDD... constants to be
defined. For some reason, these SIDD... constants aren't being defined.
Further research showed that the -msse4 switch enables support for both SSE4.1
and SSE4.2, so I tried changing the switch to -msse4, but it still errors out.
Not sure if the -msse4.2 switch automatically defines SSE4_2__ , but I tried
manually defining it, and sill no luck.
NOTE: If you want to try building it yourself on Mac, you will need to download
a different premake script file, as the included one doesn't work. You can
download the corrected script from the attachment on the second post here
https://code.google.com/p/rapidjson/issues/detail?id=54
Any ideas on how to get it building successfully on OS X ?
```
Original issue reported on code.google.com by `keith3...@rocketmail.com` on 23 Jan 2013 at 5:28
* Merged into: #55 | defect | unable to build tests on mac os x i am trying to build the tests for rapidjson on mac os x it includes three projects gtest builds fine unittest build fails and perftest build fails and when building make error out with error and error actually the first time i ran make gtest and unittest built fine but perftest failed now unittest is failing as well the compiler output shows the following errors for both unittest and perftest which causes make to fail include rapidjson reader h in function ‘const char rapidjson skipwhitespace simd const char ’ include rapidjson reader h error ‘ sidd ubyte ops’ was not declared in this scope include rapidjson reader h error ‘ sidd cmp equal any’ was not declared in this scope include rapidjson reader h error ‘ sidd bit mask’ was not declared in this scope include rapidjson reader h error ‘ sidd negative polarity’ was not declared in this scope these pre processor constants are related to instructions the makefile includes the compiler switch to enable support and looking through the header files reveal that on os x both and pre processor constants need to be defined for the sidd constants to be defined for some reason these sidd constants aren t being defined further research showed that the switch enables support for both and so i tried changing the switch to but it still errors out not sure if the switch automatically defines but i tried manually defining it and sill no luck note if you want to try building it yourself on mac you will need to download a different premake script file as the included one doesn t work you can download the corrected script from the attachment on the second post here any ideas on how to get it building successfully on os x original issue reported on code google com by rocketmail com on jan at merged into | 1 |
46,033 | 13,055,841,879 | IssuesEvent | 2020-07-30 02:53:52 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | opened | icerec/trunk dox on software.icecube.wisc.edu are "403 Forbidden" (Trac #474) | Incomplete Migration Migrated from Trac combo reconstruction defect | Migrated from https://code.icecube.wisc.edu/ticket/474
```json
{
"status": "closed",
"changetime": "2014-10-03T18:27:09",
"description": "On http://software.icecube.wisc.edu/ there is a link to IceRec \"nightly builds\": http://software.icecube.wisc.edu/icerec_trunk/ which results in \"403 - Forbidden\".\n\nI am assigning this now to dladieu, but I could equally well imagine that this is actually the job of the icerec metaproject coordinator, which would be Meike de With these days.",
"reporter": "boersma",
"cc": "meike.dewith",
"resolution": "fixed",
"_ts": "1412360829211490",
"component": "combo reconstruction",
"summary": "icerec/trunk dox on software.icecube.wisc.edu are \"403 Forbidden\"",
"priority": "minor",
"keywords": "icerec documentation doxygen",
"time": "2013-11-26T10:42:01",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
| 1.0 | icerec/trunk dox on software.icecube.wisc.edu are "403 Forbidden" (Trac #474) - Migrated from https://code.icecube.wisc.edu/ticket/474
```json
{
"status": "closed",
"changetime": "2014-10-03T18:27:09",
"description": "On http://software.icecube.wisc.edu/ there is a link to IceRec \"nightly builds\": http://software.icecube.wisc.edu/icerec_trunk/ which results in \"403 - Forbidden\".\n\nI am assigning this now to dladieu, but I could equally well imagine that this is actually the job of the icerec metaproject coordinator, which would be Meike de With these days.",
"reporter": "boersma",
"cc": "meike.dewith",
"resolution": "fixed",
"_ts": "1412360829211490",
"component": "combo reconstruction",
"summary": "icerec/trunk dox on software.icecube.wisc.edu are \"403 Forbidden\"",
"priority": "minor",
"keywords": "icerec documentation doxygen",
"time": "2013-11-26T10:42:01",
"milestone": "",
"owner": "nega",
"type": "defect"
}
```
| defect | icerec trunk dox on software icecube wisc edu are forbidden trac migrated from json status closed changetime description on there is a link to icerec nightly builds which results in forbidden n ni am assigning this now to dladieu but i could equally well imagine that this is actually the job of the icerec metaproject coordinator which would be meike de with these days reporter boersma cc meike dewith resolution fixed ts component combo reconstruction summary icerec trunk dox on software icecube wisc edu are forbidden priority minor keywords icerec documentation doxygen time milestone owner nega type defect | 1 |
36,275 | 7,875,182,247 | IssuesEvent | 2018-06-25 19:32:51 | bridgedotnet/Bridge | https://api.github.com/repos/bridgedotnet/Bridge | closed | Inheriting from an external class (no base() call case) | defect | A class-successor has to call a default constructor of an external base class even when `base()` call is not specified explicitly.
Related to #2735 and #2189.
### Steps To Reproduce
https://deck.net/abf6fe87cb3641732d977c0f8543788b
```csharp
public class Program
{
[Init(InitPosition.Top)]
public static void Init()
{
/*@
var A = (function () {
function A() {
System.Console.WriteLine("A initialized.");
}
return A;
}());
*/
}
public static void Main()
{
var b = new B();
}
}
[Virtual]
[Namespace(false)]
public class A
{
}
public class B : A
{
public B()
// : base() // Uncomment to fix
{
}
}
```
### Expected Result
Console output:
```
A initialized.
```
### Actual Result
Nothing was written to output.
| 1.0 | Inheriting from an external class (no base() call case) - A class-successor has to call a default constructor of an external base class even when `base()` call is not specified explicitly.
Related to #2735 and #2189.
### Steps To Reproduce
https://deck.net/abf6fe87cb3641732d977c0f8543788b
```csharp
public class Program
{
[Init(InitPosition.Top)]
public static void Init()
{
/*@
var A = (function () {
function A() {
System.Console.WriteLine("A initialized.");
}
return A;
}());
*/
}
public static void Main()
{
var b = new B();
}
}
[Virtual]
[Namespace(false)]
public class A
{
}
public class B : A
{
public B()
// : base() // Uncomment to fix
{
}
}
```
### Expected Result
Console output:
```
A initialized.
```
### Actual Result
Nothing was written to output.
| defect | inheriting from an external class no base call case a class successor has to call a default constructor of an external base class even when base call is not specified explicitly related to and steps to reproduce csharp public class program public static void init var a function function a system console writeline a initialized return a public static void main var b new b public class a public class b a public b base uncomment to fix expected result console output a initialized actual result nothing was written to output | 1 |
24,489 | 3,992,496,991 | IssuesEvent | 2016-05-10 02:06:50 | cakephp/cakephp | https://api.github.com/repos/cakephp/cakephp | closed | missing 'all' in cake Completion subcommands bake | console Defect | This is a (multiple allowed):
* [x] bug
* [ ] enhancement
* [ ] feature-discussion (RFC)
* CakePHP Version:
CakePHP v3.2.8
PHP : 7.0.6-1+donate.sury.org~trusty+1
### What you did
```shell
bin/cake Completion subcommands bake
```
### Expected Behavior
i'm expecting _all_ in command list
### Actual Behavior
```
[ bin/cake Completion subcommands bake master * ] 11:20
behavior cell component controller fixture form helper mailer migration migration_snapshot model plugin seed shell shell_helper task template test
```
| 1.0 | missing 'all' in cake Completion subcommands bake - This is a (multiple allowed):
* [x] bug
* [ ] enhancement
* [ ] feature-discussion (RFC)
* CakePHP Version:
CakePHP v3.2.8
PHP : 7.0.6-1+donate.sury.org~trusty+1
### What you did
```shell
bin/cake Completion subcommands bake
```
### Expected Behavior
i'm expecting _all_ in command list
### Actual Behavior
```
[ bin/cake Completion subcommands bake master * ] 11:20
behavior cell component controller fixture form helper mailer migration migration_snapshot model plugin seed shell shell_helper task template test
```
| defect | missing all in cake completion subcommands bake this is a multiple allowed bug enhancement feature discussion rfc cakephp version cakephp php donate sury org trusty what you did shell bin cake completion subcommands bake expected behavior i m expecting all in command list actual behavior behavior cell component controller fixture form helper mailer migration migration snapshot model plugin seed shell shell helper task template test | 1 |
72,441 | 24,119,803,869 | IssuesEvent | 2022-09-20 17:38:36 | SeleniumHQ/selenium | https://api.github.com/repos/SeleniumHQ/selenium | opened | [🐛 Bug]: isPromise() returning false for thenable objects | I-defect needs-triaging | ### What happened?
https://github.com/SeleniumHQ/selenium/commit/84dd6109ce692788467432ccee55f481fe49f2bc#r84520555
In the above commit, the `isPromise()` utility function was simplified to look for `typeof value === '[object Promise]'`, however this fails the function's JSDoc documentation that any object exposing a `then()` method would be considered a Promise.
My scripts that rely on `WebDriver.wait(<WebElementPromise>)` currently fail, but work after restoring the above change.
I propose that the previous functionality is returned to the `isPromise()` function to allow thenable objects to be treated as Promises.
### How can we reproduce the issue?
```shell
Pass an object with a `then()` method (for example a `WebElementPromise`) to `isPromise()`, it will return `false`.
```
### Relevant log output
```shell
n/a
```
### Operating System
Mac OS
### Selenium version
selenium-webdriver@4.4.0
### What are the browser(s) and version(s) where you see this issue?
Chrome
### What are the browser driver(s) and version(s) where you see this issue?
ChromeDriver 105.0.5195.19 (b9c217c128c16f53d12f9a02933fcfdec1bf49af-refs/branch-heads/5195@{#176})
### Are you using Selenium Grid?
_No response_ | 1.0 | [🐛 Bug]: isPromise() returning false for thenable objects - ### What happened?
https://github.com/SeleniumHQ/selenium/commit/84dd6109ce692788467432ccee55f481fe49f2bc#r84520555
In the above commit, the `isPromise()` utility function was simplified to look for `typeof value === '[object Promise]'`, however this fails the function's JSDoc documentation that any object exposing a `then()` method would be considered a Promise.
My scripts that rely on `WebDriver.wait(<WebElementPromise>)` currently fail, but work after restoring the above change.
I propose that the previous functionality is returned to the `isPromise()` function to allow thenable objects to be treated as Promises.
### How can we reproduce the issue?
```shell
Pass an object with a `then()` method (for example a `WebElementPromise`) to `isPromise()`, it will return `false`.
```
### Relevant log output
```shell
n/a
```
### Operating System
Mac OS
### Selenium version
selenium-webdriver@4.4.0
### What are the browser(s) and version(s) where you see this issue?
Chrome
### What are the browser driver(s) and version(s) where you see this issue?
ChromeDriver 105.0.5195.19 (b9c217c128c16f53d12f9a02933fcfdec1bf49af-refs/branch-heads/5195@{#176})
### Are you using Selenium Grid?
_No response_ | defect | ispromise returning false for thenable objects what happened in the above commit the ispromise utility function was simplified to look for typeof value however this fails the function s jsdoc documentation that any object exposing a then method would be considered a promise my scripts that rely on webdriver wait currently fail but work after restoring the above change i propose that the previous functionality is returned to the ispromise function to allow thenable objects to be treated as promises how can we reproduce the issue shell pass an object with a then method for example a webelementpromise to ispromise it will return false relevant log output shell n a operating system mac os selenium version selenium webdriver what are the browser s and version s where you see this issue chrome what are the browser driver s and version s where you see this issue chromedriver refs branch heads are you using selenium grid no response | 1 |
11,654 | 2,660,023,444 | IssuesEvent | 2015-03-19 01:41:49 | perfsonar/project | https://api.github.com/repos/perfsonar/project | closed | Review configuration examples in ls_registration_daemon.conf file | Component-LSRegistrationDaemon Milestone-Release3.4.2 Priority-Medium Type-Defect | Original [issue 1046](https://code.google.com/p/perfsonar-ps/issues/detail?id=1046) created by arlake228 on 2015-01-15T08:48:09.000Z:
<b>What steps will reproduce the problem?</b>
In the last suggestion from the mailing list it was written that the ls_registration_daemon.conf configuration file should have entries like:
<service_template>
<local_latency_service>
autodiscover_addresses 1
primary_interface p1p1
</local_latency_service>
<b>What is the expected output? What do you see instead?</b>
However looking at https://code.google.com/p/perfsonar-ps/source/browse/perfSONAR_PS-LSRegistrationDaemon/etc/ls_registration_daemon.conf?name=3.4
the suggested configuration syntax in comments is different:
<service_template local_latency_service>
autodiscover_addresses 1
# primary_address eth0
</service_template>
<service_template local_bandwidth_service>
autodiscover_addresses 1
# primary_address eth1
</service_template>
It seems to be wrong and/or outdated. I suggest to review the file and correct to provide correct examples for users.
| 1.0 | Review configuration examples in ls_registration_daemon.conf file - Original [issue 1046](https://code.google.com/p/perfsonar-ps/issues/detail?id=1046) created by arlake228 on 2015-01-15T08:48:09.000Z:
<b>What steps will reproduce the problem?</b>
In the last suggestion from the mailing list it was written that the ls_registration_daemon.conf configuration file should have entries like:
<service_template>
<local_latency_service>
autodiscover_addresses 1
primary_interface p1p1
</local_latency_service>
<b>What is the expected output? What do you see instead?</b>
However looking at https://code.google.com/p/perfsonar-ps/source/browse/perfSONAR_PS-LSRegistrationDaemon/etc/ls_registration_daemon.conf?name=3.4
the suggested configuration syntax in comments is different:
<service_template local_latency_service>
autodiscover_addresses 1
# primary_address eth0
</service_template>
<service_template local_bandwidth_service>
autodiscover_addresses 1
# primary_address eth1
</service_template>
It seems to be wrong and/or outdated. I suggest to review the file and correct to provide correct examples for users.
| defect | review configuration examples in ls registration daemon conf file original created by on what steps will reproduce the problem in the last suggestion from the mailing list it was written that the ls registration daemon conf configuration file should have entries like lt service template gt lt local latency service gt autodiscover addresses primary interface lt local latency service gt what is the expected output what do you see instead however looking at the suggested configuration syntax in comments is different lt service template local latency service gt autodiscover addresses primary address lt service template gt lt service template local bandwidth service gt autodiscover addresses primary address lt service template gt it seems to be wrong and or outdated i suggest to review the file and correct to provide correct examples for users | 1 |
290,201 | 25,042,595,410 | IssuesEvent | 2022-11-04 23:01:30 | godotengine/godot | https://api.github.com/repos/godotengine/godot | closed | Undo Close Tab is unreliable | bug topic:editor usability needs testing | <!-- Please search existing issues for potential duplicates before filing yours:
https://github.com/godotengine/godot/issues?q=is%3Aissue
-->
**Godot version:**
<!-- Specify commit hash if using non-official build. -->
3.2.2
**Issue description:**
<!-- What happened, and what was expected. -->
Unfortunately I don't have exact reproduction steps, but sometimes when you close scene tab and use "Undo Close Tab" (in my case with shortcut), a different tab is opened. Previously I thought that the order might get messed up, but now I think that some scenes just don't get saved in scene history and trying to reopen them reopens a different tab. This seems to happen randomly and then can be reproduced multiple times on that scene, but eventually fixes itself.
**Steps to reproduce:**
1. Open a scene
2. Open another scene
3. Close both scenes
4. Reopen scene tab
5. If you are unlucky, wrong scene gets reopened
| 1.0 | Undo Close Tab is unreliable - <!-- Please search existing issues for potential duplicates before filing yours:
https://github.com/godotengine/godot/issues?q=is%3Aissue
-->
**Godot version:**
<!-- Specify commit hash if using non-official build. -->
3.2.2
**Issue description:**
<!-- What happened, and what was expected. -->
Unfortunately I don't have exact reproduction steps, but sometimes when you close scene tab and use "Undo Close Tab" (in my case with shortcut), a different tab is opened. Previously I thought that the order might get messed up, but now I think that some scenes just don't get saved in scene history and trying to reopen them reopens a different tab. This seems to happen randomly and then can be reproduced multiple times on that scene, but eventually fixes itself.
**Steps to reproduce:**
1. Open a scene
2. Open another scene
3. Close both scenes
4. Reopen scene tab
5. If you are unlucky, wrong scene gets reopened
| non_defect | undo close tab is unreliable please search existing issues for potential duplicates before filing yours godot version issue description unfortunately i don t have exact reproduction steps but sometimes when you close scene tab and use undo close tab in my case with shortcut a different tab is opened previously i thought that the order might get messed up but now i think that some scenes just don t get saved in scene history and trying to reopen them reopens a different tab this seems to happen randomly and then can be reproduced multiple times on that scene but eventually fixes itself steps to reproduce open a scene open another scene close both scenes reopen scene tab if you are unlucky wrong scene gets reopened | 0 |
228 | 2,520,557,016 | IssuesEvent | 2015-01-19 05:07:10 | AtlasOfLivingAustralia/biocache-hubs | https://api.github.com/repos/AtlasOfLivingAustralia/biocache-hubs | closed | Fish or Fishes | priority-medium status-new type-defect |
*migrated from:* https://code.google.com/p/ala/issues/detail?id=98
*date:* Thu Aug 8 16:27:28 2013
*author:* moyesyside
---
Original Issue reported by Reported by john.t...@austmus.gov.au, Today (16 hours ago) - [https://code.google.com/p/ala-portal/issues/detail?id=300](https://code.google.com/p/ala-portal/issues/detail?id=300)
Reported by john.tann@austmus.gov.au, Today (16 hours ago)
I have been told that when we refer to 'fish' as a lifeform we should use the word 'fishes'
See [http://australianmuseum.net.au/fish-and-fishes](http://australianmuseum.net.au/fish-and-fishes)
The word 'fish' should be changed to 'fishes' when referring to a lifeform.
**What steps will reproduce the problem?**
An example:
Australian Museum Ichthyology collection. in the graph of records by higher level group:
Axis is Arthropods Fish Insects Reptiles
**What is the expected output? What do you see instead?**
Axis is Arthropods ***Fishes*** Insects Reptiles
What version of the product are you using? On what operating system?
2013
Please provide any additional information below.
The term 'fish' is probably incorrectly used in a bunch of places.
This is a low priority fix.
fish or fishes.png 4.7 KB View Download
| 1.0 | Fish or Fishes -
*migrated from:* https://code.google.com/p/ala/issues/detail?id=98
*date:* Thu Aug 8 16:27:28 2013
*author:* moyesyside
---
Original Issue reported by Reported by john.t...@austmus.gov.au, Today (16 hours ago) - [https://code.google.com/p/ala-portal/issues/detail?id=300](https://code.google.com/p/ala-portal/issues/detail?id=300)
Reported by john.tann@austmus.gov.au, Today (16 hours ago)
I have been told that when we refer to 'fish' as a lifeform we should use the word 'fishes'
See [http://australianmuseum.net.au/fish-and-fishes](http://australianmuseum.net.au/fish-and-fishes)
The word 'fish' should be changed to 'fishes' when referring to a lifeform.
**What steps will reproduce the problem?**
An example:
Australian Museum Ichthyology collection. in the graph of records by higher level group:
Axis is Arthropods Fish Insects Reptiles
**What is the expected output? What do you see instead?**
Axis is Arthropods ***Fishes*** Insects Reptiles
What version of the product are you using? On what operating system?
2013
Please provide any additional information below.
The term 'fish' is probably incorrectly used in a bunch of places.
This is a low priority fix.
fish or fishes.png 4.7 KB View Download
| defect | fish or fishes migrated from date thu aug author moyesyside original issue reported by reported by john t austmus gov au today hours ago reported by john tann austmus gov au today hours ago i have been told that when we refer to fish as a lifeform we should use the word fishes see the word fish should be changed to fishes when referring to a lifeform what steps will reproduce the problem an example australian museum ichthyology collection in the graph of records by higher level group axis is arthropods fish insects reptiles what is the expected output what do you see instead axis is arthropods fishes insects reptiles what version of the product are you using on what operating system please provide any additional information below the term fish is probably incorrectly used in a bunch of places this is a low priority fix fish or fishes png kb view download | 1 |
83,031 | 10,318,815,523 | IssuesEvent | 2019-08-30 15:49:04 | CICE-Consortium/CICE | https://api.github.com/repos/CICE-Consortium/CICE | closed | evp kernel version 2 testing and validation | Documentation Dynamics Priority: High Testing Type: Feature | We are going to merge PR #278, PR #252. There are several outstanding issues, basically copied from the end of #252,
---------------------
Let me summarize where we are.
With evp_kernel_ver=0, results are bit-for-bit for most tests against the current master. This is running full test suites on gordon for 4 compilers. A subset of box tests are NOT bit-for-bit on 3/4 compilers. Rerunning the failed box tests with the debug flag (reduced optimization and run time checks) on both master and this PR results in bit-for-bit identical answers. It seems the changes in the answers in the box test is caused by some compiler optimization as a results of the code changes. This might be associated with the evp kernel changes (although @mhrib makes a case it shouldn't) or it might be associated with some of the code cleanup. We could look into this further or we could accept it. Personally, I am comfortable with this outcome as it stands. I believe we've shown the answers are roundoff different (see above gbox128 diff) as a result of compiler optimization and that we can make this bit-for-bit if we reduce compiler optimization. I think based on these results, we could merge this PR. evp_kernel_ver=0 will be the default setting.
Separately, there is an effort to test and validate the evp_kernel_ver=2. The same test suite on gordon was run with the new kernel on. Results can be found https://github.com/CICE-Consortium/Test-Results/wiki/cice_by_hash_forks, hash aa6de33...+evpk=2. Three to four tests fail on each compiler, and they are the same tests across the compilers. Looking at the intel results, https://github.com/CICE-Consortium/Test-Results/wiki/aa6de33f19.gordon.pgi.190128.235649, there are four failures.
- restart gbox128 4x2. This test runs but fails to restart exactly.
- restart gx1 40x4 droundrobin medium. This test fails with "(abort_ice) error = (horizontal_remap)ERROR: bad departure points" on the first timestep.
- restart gx3 16x2x5x10x20 drakeX2. This test fails with "(abort_ice) error = (horizontal_remap)ERROR: bad departure points" on the first timestep.
- restart tx1 40x4 dsectrobin medium. This test fails gracefully in the evp kernel. tx1 is not supported yet.
Again, many tests passed, but these 4 failures need to be debugged. In addition, the qc test relies on the gx1 configuration, so the qc testing comparing evp_kernel_ver=2 to 0 could not be done.
So, the outstanding tasks are
- debug the 4 failures noted above
- run the qc test comparing evp_kernel_ver=0 to evp_kernel_ver=2. This requires gx1 (one of the failing tests)
- update documentation
- change evp_kernel_ver variable to kevp_kernel
- produce and document timing information comparing evp_kernel_ver=0 and 2.
- add evp_kernel_ver=2 tests to the test suite
- maybe do a little cleanup on ice_dyn_evp_1d.F90 to make the code a little more readable (breaks between subroutines and such)
| 1.0 | evp kernel version 2 testing and validation - We are going to merge PR #278, PR #252. There are several outstanding issues, basically copied from the end of #252,
---------------------
Let me summarize where we are.
With evp_kernel_ver=0, results are bit-for-bit for most tests against the current master. This is running full test suites on gordon for 4 compilers. A subset of box tests are NOT bit-for-bit on 3/4 compilers. Rerunning the failed box tests with the debug flag (reduced optimization and run time checks) on both master and this PR results in bit-for-bit identical answers. It seems the changes in the answers in the box test is caused by some compiler optimization as a results of the code changes. This might be associated with the evp kernel changes (although @mhrib makes a case it shouldn't) or it might be associated with some of the code cleanup. We could look into this further or we could accept it. Personally, I am comfortable with this outcome as it stands. I believe we've shown the answers are roundoff different (see above gbox128 diff) as a result of compiler optimization and that we can make this bit-for-bit if we reduce compiler optimization. I think based on these results, we could merge this PR. evp_kernel_ver=0 will be the default setting.
Separately, there is an effort to test and validate the evp_kernel_ver=2. The same test suite on gordon was run with the new kernel on. Results can be found https://github.com/CICE-Consortium/Test-Results/wiki/cice_by_hash_forks, hash aa6de33...+evpk=2. Three to four tests fail on each compiler, and they are the same tests across the compilers. Looking at the intel results, https://github.com/CICE-Consortium/Test-Results/wiki/aa6de33f19.gordon.pgi.190128.235649, there are four failures.
- restart gbox128 4x2. This test runs but fails to restart exactly.
- restart gx1 40x4 droundrobin medium. This test fails with "(abort_ice) error = (horizontal_remap)ERROR: bad departure points" on the first timestep.
- restart gx3 16x2x5x10x20 drakeX2. This test fails with "(abort_ice) error = (horizontal_remap)ERROR: bad departure points" on the first timestep.
- restart tx1 40x4 dsectrobin medium. This test fails gracefully in the evp kernel. tx1 is not supported yet.
Again, many tests passed, but these 4 failures need to be debugged. In addition, the qc test relies on the gx1 configuration, so the qc testing comparing evp_kernel_ver=2 to 0 could not be done.
So, the outstanding tasks are
- debug the 4 failures noted above
- run the qc test comparing evp_kernel_ver=0 to evp_kernel_ver=2. This requires gx1 (one of the failing tests)
- update documentation
- change evp_kernel_ver variable to kevp_kernel
- produce and document timing information comparing evp_kernel_ver=0 and 2.
- add evp_kernel_ver=2 tests to the test suite
- maybe do a little cleanup on ice_dyn_evp_1d.F90 to make the code a little more readable (breaks between subroutines and such)
| non_defect | evp kernel version testing and validation we are going to merge pr pr there are several outstanding issues basically copied from the end of let me summarize where we are with evp kernel ver results are bit for bit for most tests against the current master this is running full test suites on gordon for compilers a subset of box tests are not bit for bit on compilers rerunning the failed box tests with the debug flag reduced optimization and run time checks on both master and this pr results in bit for bit identical answers it seems the changes in the answers in the box test is caused by some compiler optimization as a results of the code changes this might be associated with the evp kernel changes although mhrib makes a case it shouldn t or it might be associated with some of the code cleanup we could look into this further or we could accept it personally i am comfortable with this outcome as it stands i believe we ve shown the answers are roundoff different see above diff as a result of compiler optimization and that we can make this bit for bit if we reduce compiler optimization i think based on these results we could merge this pr evp kernel ver will be the default setting separately there is an effort to test and validate the evp kernel ver the same test suite on gordon was run with the new kernel on results can be found hash evpk three to four tests fail on each compiler and they are the same tests across the compilers looking at the intel results there are four failures restart this test runs but fails to restart exactly restart droundrobin medium this test fails with abort ice error horizontal remap error bad departure points on the first timestep restart this test fails with abort ice error horizontal remap error bad departure points on the first timestep restart dsectrobin medium this test fails gracefully in the evp kernel is not supported yet again many tests passed but these failures need to be debugged in addition the qc test relies on the configuration so the qc testing comparing evp kernel ver to could not be done so the outstanding tasks are debug the failures noted above run the qc test comparing evp kernel ver to evp kernel ver this requires one of the failing tests update documentation change evp kernel ver variable to kevp kernel produce and document timing information comparing evp kernel ver and add evp kernel ver tests to the test suite maybe do a little cleanup on ice dyn evp to make the code a little more readable breaks between subroutines and such | 0 |
27,811 | 5,106,884,223 | IssuesEvent | 2017-01-05 13:10:36 | TASVideos/BizHawk | https://api.github.com/repos/TASVideos/BizHawk | closed | Genplus-gx [BizHawk] Game Genie codes not working | bug Core-EmuHawk Core-Genplus-GX Core-GensHawk OpSys-Any Priority-Medium Type-Defect | I can't get any Game Genie codes to work on BizHawk Genplus-gx. I have codes that I know that work but on BizHawk they don't work. I know how to use BizHawk. Like on the game Strider (UE) [!]. Here is a code that give you Infinite Life In The Life Gauge.
AK8T-AA5R
AKVT-AA94
AJLT-AA9E
On Kega this code works.
| 1.0 | Genplus-gx [BizHawk] Game Genie codes not working - I can't get any Game Genie codes to work on BizHawk Genplus-gx. I have codes that I know that work but on BizHawk they don't work. I know how to use BizHawk. Like on the game Strider (UE) [!]. Here is a code that give you Infinite Life In The Life Gauge.
AK8T-AA5R
AKVT-AA94
AJLT-AA9E
On Kega this code works.
| defect | genplus gx game genie codes not working i can t get any game genie codes to work on bizhawk genplus gx i have codes that i know that work but on bizhawk they don t work i know how to use bizhawk like on the game strider ue here is a code that give you infinite life in the life gauge akvt ajlt on kega this code works | 1 |
277,173 | 21,016,821,004 | IssuesEvent | 2022-03-30 11:49:32 | SW-Team2/se-team2-tetris | https://api.github.com/repos/SW-Team2/se-team2-tetris | closed | docs: PR, Issue template 제작 | 📄 documentation | ## 개요
- PR과 Issue을 보면서 명확하게 어떤 것을 했는지 전달이 안되는 부분이 있었습니다.
- 예를 들어, 어떤 기능을 구현했을 때, 이것이 어떻게 동작하는지 캡쳐 화면이라던가 동영상 등 추가적인 정보가 없어 직접 실행해야 하는 수고로움이 있었습니다.
- 탬플릿을 제공하여 다른 사람에게 명확하게 정보를 제공하고자 합니다.
## 할 일
- [x] PR template 제작
- [x] Issue template 제작
## 참고
https://docs.github.com/en/communities/using-templates-to-encourage-useful-issues-and-pull-requests/about-issue-and-pull-request-templates
| 1.0 | docs: PR, Issue template 제작 - ## 개요
- PR과 Issue을 보면서 명확하게 어떤 것을 했는지 전달이 안되는 부분이 있었습니다.
- 예를 들어, 어떤 기능을 구현했을 때, 이것이 어떻게 동작하는지 캡쳐 화면이라던가 동영상 등 추가적인 정보가 없어 직접 실행해야 하는 수고로움이 있었습니다.
- 탬플릿을 제공하여 다른 사람에게 명확하게 정보를 제공하고자 합니다.
## 할 일
- [x] PR template 제작
- [x] Issue template 제작
## 참고
https://docs.github.com/en/communities/using-templates-to-encourage-useful-issues-and-pull-requests/about-issue-and-pull-request-templates
| non_defect | docs pr issue template 제작 개요 pr과 issue을 보면서 명확하게 어떤 것을 했는지 전달이 안되는 부분이 있었습니다 예를 들어 어떤 기능을 구현했을 때 이것이 어떻게 동작하는지 캡쳐 화면이라던가 동영상 등 추가적인 정보가 없어 직접 실행해야 하는 수고로움이 있었습니다 탬플릿을 제공하여 다른 사람에게 명확하게 정보를 제공하고자 합니다 할 일 pr template 제작 issue template 제작 참고 | 0 |
225,023 | 17,789,594,198 | IssuesEvent | 2021-08-31 14:48:40 | elastic/kibana | https://api.github.com/repos/elastic/kibana | opened | [RAC] [Observability] Create functional tests for alerts table add-to-case actions | test Team:logs-metrics-ui Theme: rac | ## :notebook: Summary
We should create functional tests for the add-to-case actions on the Observability Alerts page.
:link: makes use of the service created in #110627
## :heavy_check_mark: Test cases
:warning: When you pick this up, please spend a minute to think about and expand on these test cases. This is just an initial proposal.
- add-to-case row actions are...
- enabled with `all` permissions for the different observability solutions owning the respective rules
- enabled for APM alerts with special APM alert permissions
- disabled with `read` and `none` permissions for the different observability solutions owning the respective rules | 1.0 | [RAC] [Observability] Create functional tests for alerts table add-to-case actions - ## :notebook: Summary
We should create functional tests for the add-to-case actions on the Observability Alerts page.
:link: makes use of the service created in #110627
## :heavy_check_mark: Test cases
:warning: When you pick this up, please spend a minute to think about and expand on these test cases. This is just an initial proposal.
- add-to-case row actions are...
- enabled with `all` permissions for the different observability solutions owning the respective rules
- enabled for APM alerts with special APM alert permissions
- disabled with `read` and `none` permissions for the different observability solutions owning the respective rules | non_defect | create functional tests for alerts table add to case actions notebook summary we should create functional tests for the add to case actions on the observability alerts page link makes use of the service created in heavy check mark test cases warning when you pick this up please spend a minute to think about and expand on these test cases this is just an initial proposal add to case row actions are enabled with all permissions for the different observability solutions owning the respective rules enabled for apm alerts with special apm alert permissions disabled with read and none permissions for the different observability solutions owning the respective rules | 0 |
46,225 | 13,055,872,553 | IssuesEvent | 2020-07-30 02:59:16 | icecube-trac/tix2 | https://api.github.com/repos/icecube-trac/tix2 | opened | Matplotlib crashes on systems with AMD bulldozer architecture. (Trac #795) | Incomplete Migration Migrated from Trac defect tools/ports | Migrated from https://code.icecube.wisc.edu/ticket/795
```json
{
"status": "closed",
"changetime": "2014-10-30T18:25:07",
"description": "Matplotlib provided by cvmfs crashes on SL6 with AMD bulldozer processors (AMD FX-8350).\n{{{\npython -c \"import pylab; pylab.plot(1,2); pylab.show()\"\n}}}\n\n\ncrashes with the message\n{{{ \nIllegal instruction\n}}}\n\nGDB hints at libgoto as the source of the problem (see attachment)\n",
"reporter": "chaack",
"cc": "nega",
"resolution": "fixed",
"_ts": "1414693507102825",
"component": "tools/ports",
"summary": "Matplotlib crashes on systems with AMD bulldozer architecture.",
"priority": "major",
"keywords": "BLAS cvmfs",
"time": "2014-10-25T14:34:25",
"milestone": "",
"owner": "david.schultz",
"type": "defect"
}
```
| 1.0 | Matplotlib crashes on systems with AMD bulldozer architecture. (Trac #795) - Migrated from https://code.icecube.wisc.edu/ticket/795
```json
{
"status": "closed",
"changetime": "2014-10-30T18:25:07",
"description": "Matplotlib provided by cvmfs crashes on SL6 with AMD bulldozer processors (AMD FX-8350).\n{{{\npython -c \"import pylab; pylab.plot(1,2); pylab.show()\"\n}}}\n\n\ncrashes with the message\n{{{ \nIllegal instruction\n}}}\n\nGDB hints at libgoto as the source of the problem (see attachment)\n",
"reporter": "chaack",
"cc": "nega",
"resolution": "fixed",
"_ts": "1414693507102825",
"component": "tools/ports",
"summary": "Matplotlib crashes on systems with AMD bulldozer architecture.",
"priority": "major",
"keywords": "BLAS cvmfs",
"time": "2014-10-25T14:34:25",
"milestone": "",
"owner": "david.schultz",
"type": "defect"
}
```
| defect | matplotlib crashes on systems with amd bulldozer architecture trac migrated from json status closed changetime description matplotlib provided by cvmfs crashes on with amd bulldozer processors amd fx n npython c import pylab pylab plot pylab show n n n ncrashes with the message n nillegal instruction n n ngdb hints at libgoto as the source of the problem see attachment n reporter chaack cc nega resolution fixed ts component tools ports summary matplotlib crashes on systems with amd bulldozer architecture priority major keywords blas cvmfs time milestone owner david schultz type defect | 1 |
9,750 | 2,615,167,204 | IssuesEvent | 2015-03-01 06:48:01 | chrsmith/reaver-wps | https://api.github.com/repos/chrsmith/reaver-wps | opened | 99.99% Reaver repeats same pin witn no m5 packets | auto-migrated Priority-Triage Type-Defect | ```
0. What version of Reaver are you using? (Only defects against the latest
version will be considered.)
, reaver 1.4
1. What operating system are you using (Linux is the only supported OS)?
Backtrack 5 r3 usb boot
2. Is your wireless card in monitor mode (yes/no)?
yes
3. What is the signal strength of the Access Point you are trying to crack?
-60
4. What is the manufacturer and model # of the device you are trying to
crack?
BtHub4
5. What is the entire command line string you are supplying to reaver?
reaver -i mon0 -b CC:BB:33:A3:BD:7C -d 0 -t 0 -N -S
6. Please describe what you think the issue is.
Hi everyone. For some reason I run an attack against an AP and I got to 99.99%
with pin tstarting with 998 well i mean what are the chances that the password
is not in the range of 0 to 998 .... and no m5 packets are sent. or recieved
therefore I assume the first part of the pin was incorrect... If so then reaver
missed the correct pin? I recieved Warning: Timeout and 0.02 and 0.03 error,
however the cracking progressed... Also I noticed that when crakcking reaver
does this:
Trying pin 23456789
Trying pin 23457780
Trying pin 23458781
Something like this, therefore Im not sure but is it possible that reaver
missed the correct pin??
7. Paste the output from Reaver below.
Sadly I do not have it. I'm writing from Windows and I failed to record the log
however
m1 packets
m2 packet
m3 packet
m4 packet
Wsc Nack
Eapol REquest
Trying pin XXXXXXX
^repeat....
and then it goes to 99.99% and it took me 20 hours to do it.. repeats the same
pin over and over which start with 998.......... chances are so low......
```
Original issue reported on code.google.com by `Krer...@gmail.com` on 12 Nov 2013 at 12:15 | 1.0 | 99.99% Reaver repeats same pin witn no m5 packets - ```
0. What version of Reaver are you using? (Only defects against the latest
version will be considered.)
, reaver 1.4
1. What operating system are you using (Linux is the only supported OS)?
Backtrack 5 r3 usb boot
2. Is your wireless card in monitor mode (yes/no)?
yes
3. What is the signal strength of the Access Point you are trying to crack?
-60
4. What is the manufacturer and model # of the device you are trying to
crack?
BtHub4
5. What is the entire command line string you are supplying to reaver?
reaver -i mon0 -b CC:BB:33:A3:BD:7C -d 0 -t 0 -N -S
6. Please describe what you think the issue is.
Hi everyone. For some reason I run an attack against an AP and I got to 99.99%
with pin tstarting with 998 well i mean what are the chances that the password
is not in the range of 0 to 998 .... and no m5 packets are sent. or recieved
therefore I assume the first part of the pin was incorrect... If so then reaver
missed the correct pin? I recieved Warning: Timeout and 0.02 and 0.03 error,
however the cracking progressed... Also I noticed that when crakcking reaver
does this:
Trying pin 23456789
Trying pin 23457780
Trying pin 23458781
Something like this, therefore Im not sure but is it possible that reaver
missed the correct pin??
7. Paste the output from Reaver below.
Sadly I do not have it. I'm writing from Windows and I failed to record the log
however
m1 packets
m2 packet
m3 packet
m4 packet
Wsc Nack
Eapol REquest
Trying pin XXXXXXX
^repeat....
and then it goes to 99.99% and it took me 20 hours to do it.. repeats the same
pin over and over which start with 998.......... chances are so low......
```
Original issue reported on code.google.com by `Krer...@gmail.com` on 12 Nov 2013 at 12:15 | defect | reaver repeats same pin witn no packets what version of reaver are you using only defects against the latest version will be considered reaver what operating system are you using linux is the only supported os backtrack usb boot is your wireless card in monitor mode yes no yes what is the signal strength of the access point you are trying to crack what is the manufacturer and model of the device you are trying to crack what is the entire command line string you are supplying to reaver reaver i b cc bb bd d t n s please describe what you think the issue is hi everyone for some reason i run an attack against an ap and i got to with pin tstarting with well i mean what are the chances that the password is not in the range of to and no packets are sent or recieved therefore i assume the first part of the pin was incorrect if so then reaver missed the correct pin i recieved warning timeout and and error however the cracking progressed also i noticed that when crakcking reaver does this trying pin trying pin trying pin something like this therefore im not sure but is it possible that reaver missed the correct pin paste the output from reaver below sadly i do not have it i m writing from windows and i failed to record the log however packets packet packet packet wsc nack eapol request trying pin xxxxxxx repeat and then it goes to and it took me hours to do it repeats the same pin over and over which start with chances are so low original issue reported on code google com by krer gmail com on nov at | 1 |
440,647 | 30,754,420,055 | IssuesEvent | 2023-07-28 23:34:16 | aws-samples/amazon-kinesis-video-streams-demos | https://api.github.com/repos/aws-samples/amazon-kinesis-video-streams-demos | closed | Pass AWS and stream link arguments in docker run | documentation question | How can I pass the:
`AWS_ACCESS_KEY_ID=<AWS_ACCESS_KEY_ID> AWS_SECRET_ACCESS_KEY=<AWS_SECRET_ACCESS_KEY> ./kvs_gstreamer_sample <STREAM_NAME> <RTSP_URL> `
Directly in the docker run? | 1.0 | Pass AWS and stream link arguments in docker run - How can I pass the:
`AWS_ACCESS_KEY_ID=<AWS_ACCESS_KEY_ID> AWS_SECRET_ACCESS_KEY=<AWS_SECRET_ACCESS_KEY> ./kvs_gstreamer_sample <STREAM_NAME> <RTSP_URL> `
Directly in the docker run? | non_defect | pass aws and stream link arguments in docker run how can i pass the aws access key id aws secret access key kvs gstreamer sample directly in the docker run | 0 |
23,541 | 4,021,604,118 | IssuesEvent | 2016-05-16 22:40:30 | elastic/logstash | https://api.github.com/repos/elastic/logstash | closed | Hot Threads failure in Travis CI | bug tests | Using our new [Travis CI branch](https://github.com/elastic/logstash/pull/4608) we sometimes get [this odd failure](https://travis-ci.org/elastic/logstash/builds/124009233).
This seems like an intermittently failing test we'll have to look into.
@purbon do you understand what's going on here? | 1.0 | Hot Threads failure in Travis CI - Using our new [Travis CI branch](https://github.com/elastic/logstash/pull/4608) we sometimes get [this odd failure](https://travis-ci.org/elastic/logstash/builds/124009233).
This seems like an intermittently failing test we'll have to look into.
@purbon do you understand what's going on here? | non_defect | hot threads failure in travis ci using our new we sometimes get this seems like an intermittently failing test we ll have to look into purbon do you understand what s going on here | 0 |
43,241 | 23,163,299,355 | IssuesEvent | 2022-07-29 20:28:01 | mattermost/focalboard | https://api.github.com/repos/mattermost/focalboard | closed | PERF: Fetch board members in parallel | Bug Sev/1 Performance | Here in web app codebase, we fetch each individual board member's details in a separate API call - https://github.com/mattermost/focalboard/blob/3c7fd72dcf0456e14b324cfb0621371289603f2a/webapp/src/store/boards.ts#L36
I propose fetching users in bulk instead of one by one. | True | PERF: Fetch board members in parallel - Here in web app codebase, we fetch each individual board member's details in a separate API call - https://github.com/mattermost/focalboard/blob/3c7fd72dcf0456e14b324cfb0621371289603f2a/webapp/src/store/boards.ts#L36
I propose fetching users in bulk instead of one by one. | non_defect | perf fetch board members in parallel here in web app codebase we fetch each individual board member s details in a separate api call i propose fetching users in bulk instead of one by one | 0 |
489,843 | 14,112,715,084 | IssuesEvent | 2020-11-07 07:04:53 | AY2021S1-CS2103T-W12-4/tp | https://api.github.com/repos/AY2021S1-CS2103T-W12-4/tp | closed | Update UG - Consistent examples across UG and code | priority.High type.Task type.UserGuide | TODOS:
1. Update command summary section
2. Change parameter name for students from `NAME` to `STUDENT_NAME`
3. For individual commands, update to the latest examples (ref google docs)
4. For commands with more than 1 method, label each of the formats to the respective method (ie `markpresent`)
| 1.0 | Update UG - Consistent examples across UG and code - TODOS:
1. Update command summary section
2. Change parameter name for students from `NAME` to `STUDENT_NAME`
3. For individual commands, update to the latest examples (ref google docs)
4. For commands with more than 1 method, label each of the formats to the respective method (ie `markpresent`)
| non_defect | update ug consistent examples across ug and code todos update command summary section change parameter name for students from name to student name for individual commands update to the latest examples ref google docs for commands with more than method label each of the formats to the respective method ie markpresent | 0 |
361,307 | 10,707,104,460 | IssuesEvent | 2019-10-24 16:43:11 | netdata/netdata | https://api.github.com/repos/netdata/netdata | closed | Generate alarms if the disk cannot keep up with data collection | area/database feature request priority/high | <!---
When creating a feature request please:
- Verify first that your issue is not already reported on GitHub
- Explain new feature briefly in "Feature idea summary" section
- Provide a clear and concise description of what you expect to happen.
--->
##### Feature idea summary
If the disk cannot keep up with metric collection this condition should be detectable and alarms should be generated. This is for `memory mode = dbengine`.
| 1.0 | Generate alarms if the disk cannot keep up with data collection - <!---
When creating a feature request please:
- Verify first that your issue is not already reported on GitHub
- Explain new feature briefly in "Feature idea summary" section
- Provide a clear and concise description of what you expect to happen.
--->
##### Feature idea summary
If the disk cannot keep up with metric collection this condition should be detectable and alarms should be generated. This is for `memory mode = dbengine`.
| non_defect | generate alarms if the disk cannot keep up with data collection when creating a feature request please verify first that your issue is not already reported on github explain new feature briefly in feature idea summary section provide a clear and concise description of what you expect to happen feature idea summary if the disk cannot keep up with metric collection this condition should be detectable and alarms should be generated this is for memory mode dbengine | 0 |
46,383 | 5,806,359,928 | IssuesEvent | 2017-05-04 02:19:00 | NSW-OEH-EMS-KST/grid-garage-3 | https://api.github.com/repos/NSW-OEH-EMS-KST/grid-garage-3 | closed | raster - tweak values | tested and working | Let users know that the min and max values must be integers.. @byezy this still the case? | 1.0 | raster - tweak values - Let users know that the min and max values must be integers.. @byezy this still the case? | non_defect | raster tweak values let users know that the min and max values must be integers byezy this still the case | 0 |
70,673 | 23,282,203,022 | IssuesEvent | 2022-08-05 13:11:52 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | Error configuring hazelcast with DSL style "IMap eviction config doesn't support max size policy `ENTRY_COUNT`" | Type: Defect | <!--
Thanks for reporting your issue. Please share with us the following information, to help us resolve your issue quickly and efficiently.
-->
**Describe the bug**
When i configure a eviction in a map (please see shared code) , once the map is getting used the app throws error (see exception below)
**Expected behavior**
I expect to get the map
**To Reproduce**
Steps to reproduce the behavior:
1. Setup a map with the eviction LRU or LFU or NONE.
2. Use by the map
3. Error is thrown
4. The app is in clustered but for this case is just a single instance.
**Additional context**
<!--
Add any other context about the problem here.
Common details that we're often interested in:
- Detailed description of the steps to reproduce your issue
- Logs and stack traces, if available
- Hazelcast version that you use (e.g. 3.4, also specify whether it is a minor release or the latest snapshot)
- If available, integration module versions (e.g. Tomcat, Jetty, Spring, Hibernate). Also, include their detailed configuration information such as web.xml, Hibernate configuration and `context.xml` for Spring
- Cluster size, i.e. the number of Hazelcast cluster members
- Number of the clients
- Version of Java. It is also helpful to mention the JVM parameters
- Operating system. If it is Linux, kernel version is helpful
- Unit test with the `hazelcast.xml` file. If you could include a unit test which reproduces your issue, we would be grateful
-->
-The application is a spring boot app.
-I'm using spring-hazelcast dependency that aims to use 5.1 version of hazelcast core.
-I tried updating to 5.1.2 , same behavior.
-Spring boot version is 2.6.8
Config config = new XmlConfigBuilder().build();
config.setClusterName("cluster-name");
config
.setInstanceName(instanceName)
.addMapConfig(new MapConfig()
.setBackupCount(1)
.setName("example-map")
.setEvictionConfig(new EvictionConfig().setEvictionPolicy(LFU))
.setMapStoreConfig(new MapStoreConfig()
.setInitialLoadMode(EAGER)
.setWriteBatchSize(500)
.setWriteCoalescing(true)
.setEnabled(true)
.setFactoryImplementation(mapStoreConfig)
.setWriteDelaySeconds(5)));
Logs:
`com.hazelcast.config.InvalidConfigurationException: IMap eviction config doesn't support max size policy `ENTRY_COUNT`. Please select a valid one: [PER_NODE, PER_PARTITION, USED_HEAP_PERCENTAGE, USED_HEAP_SIZE, FREE_HEAP_PERCENTAGE, FREE_HEAP_SIZE, USED_NATIVE_MEMORY_SIZE, USED_NATIVE_MEMORY_PERCENTAGE, FREE_NATIVE_MEMORY_SIZE, FREE_NATIVE_MEMORY_PERCENTAGE].
at com.hazelcast.internal.config.ConfigValidator.checkMapMaxSizePolicyConfig(ConfigValidator.java:199)
at com.hazelcast.internal.config.ConfigValidator.checkMapEvictionConfig(ConfigValidator.java:177)`
| 1.0 | Error configuring hazelcast with DSL style "IMap eviction config doesn't support max size policy `ENTRY_COUNT`" - <!--
Thanks for reporting your issue. Please share with us the following information, to help us resolve your issue quickly and efficiently.
-->
**Describe the bug**
When i configure a eviction in a map (please see shared code) , once the map is getting used the app throws error (see exception below)
**Expected behavior**
I expect to get the map
**To Reproduce**
Steps to reproduce the behavior:
1. Setup a map with the eviction LRU or LFU or NONE.
2. Use by the map
3. Error is thrown
4. The app is in clustered but for this case is just a single instance.
**Additional context**
<!--
Add any other context about the problem here.
Common details that we're often interested in:
- Detailed description of the steps to reproduce your issue
- Logs and stack traces, if available
- Hazelcast version that you use (e.g. 3.4, also specify whether it is a minor release or the latest snapshot)
- If available, integration module versions (e.g. Tomcat, Jetty, Spring, Hibernate). Also, include their detailed configuration information such as web.xml, Hibernate configuration and `context.xml` for Spring
- Cluster size, i.e. the number of Hazelcast cluster members
- Number of the clients
- Version of Java. It is also helpful to mention the JVM parameters
- Operating system. If it is Linux, kernel version is helpful
- Unit test with the `hazelcast.xml` file. If you could include a unit test which reproduces your issue, we would be grateful
-->
-The application is a spring boot app.
-I'm using spring-hazelcast dependency that aims to use 5.1 version of hazelcast core.
-I tried updating to 5.1.2 , same behavior.
-Spring boot version is 2.6.8
Config config = new XmlConfigBuilder().build();
config.setClusterName("cluster-name");
config
.setInstanceName(instanceName)
.addMapConfig(new MapConfig()
.setBackupCount(1)
.setName("example-map")
.setEvictionConfig(new EvictionConfig().setEvictionPolicy(LFU))
.setMapStoreConfig(new MapStoreConfig()
.setInitialLoadMode(EAGER)
.setWriteBatchSize(500)
.setWriteCoalescing(true)
.setEnabled(true)
.setFactoryImplementation(mapStoreConfig)
.setWriteDelaySeconds(5)));
Logs:
`com.hazelcast.config.InvalidConfigurationException: IMap eviction config doesn't support max size policy `ENTRY_COUNT`. Please select a valid one: [PER_NODE, PER_PARTITION, USED_HEAP_PERCENTAGE, USED_HEAP_SIZE, FREE_HEAP_PERCENTAGE, FREE_HEAP_SIZE, USED_NATIVE_MEMORY_SIZE, USED_NATIVE_MEMORY_PERCENTAGE, FREE_NATIVE_MEMORY_SIZE, FREE_NATIVE_MEMORY_PERCENTAGE].
at com.hazelcast.internal.config.ConfigValidator.checkMapMaxSizePolicyConfig(ConfigValidator.java:199)
at com.hazelcast.internal.config.ConfigValidator.checkMapEvictionConfig(ConfigValidator.java:177)`
| defect | error configuring hazelcast with dsl style imap eviction config doesn t support max size policy entry count thanks for reporting your issue please share with us the following information to help us resolve your issue quickly and efficiently describe the bug when i configure a eviction in a map please see shared code once the map is getting used the app throws error see exception below expected behavior i expect to get the map to reproduce steps to reproduce the behavior setup a map with the eviction lru or lfu or none use by the map error is thrown the app is in clustered but for this case is just a single instance additional context add any other context about the problem here common details that we re often interested in detailed description of the steps to reproduce your issue logs and stack traces if available hazelcast version that you use e g also specify whether it is a minor release or the latest snapshot if available integration module versions e g tomcat jetty spring hibernate also include their detailed configuration information such as web xml hibernate configuration and context xml for spring cluster size i e the number of hazelcast cluster members number of the clients version of java it is also helpful to mention the jvm parameters operating system if it is linux kernel version is helpful unit test with the hazelcast xml file if you could include a unit test which reproduces your issue we would be grateful the application is a spring boot app i m using spring hazelcast dependency that aims to use version of hazelcast core i tried updating to same behavior spring boot version is config config new xmlconfigbuilder build config setclustername cluster name config setinstancename instancename addmapconfig new mapconfig setbackupcount setname example map setevictionconfig new evictionconfig setevictionpolicy lfu setmapstoreconfig new mapstoreconfig setinitialloadmode eager setwritebatchsize setwritecoalescing true setenabled true setfactoryimplementation mapstoreconfig setwritedelayseconds logs com hazelcast config invalidconfigurationexception imap eviction config doesn t support max size policy entry count please select a valid one at com hazelcast internal config configvalidator checkmapmaxsizepolicyconfig configvalidator java at com hazelcast internal config configvalidator checkmapevictionconfig configvalidator java | 1 |
27,094 | 4,875,252,256 | IssuesEvent | 2016-11-16 08:57:56 | TNGSB/eWallet | https://api.github.com/repos/TNGSB/eWallet | closed | eWallet_MobileApp(Airtime)_Both IOS & Android #097 | Defect - Medium (Sev-3) | [Defect_Mobile App #97.xlsx](https://github.com/TNGSB/eWallet/files/591389/Defect_Mobile.App.97.xlsx)
Test Description : To verify the error message when user left all the fields in blank
Defect Description : System displayed wrong error message when user left all the fields in blank - apply to both IOS and Android
Refer attachment for POT | 1.0 | eWallet_MobileApp(Airtime)_Both IOS & Android #097 - [Defect_Mobile App #97.xlsx](https://github.com/TNGSB/eWallet/files/591389/Defect_Mobile.App.97.xlsx)
Test Description : To verify the error message when user left all the fields in blank
Defect Description : System displayed wrong error message when user left all the fields in blank - apply to both IOS and Android
Refer attachment for POT | defect | ewallet mobileapp airtime both ios android test description to verify the error message when user left all the fields in blank defect description system displayed wrong error message when user left all the fields in blank apply to both ios and android refer attachment for pot | 1 |
594,799 | 18,054,428,399 | IssuesEvent | 2021-09-20 05:48:22 | naev/naev | https://api.github.com/repos/naev/naev | closed | mission markers are hard to distinguish | Type-Enhancement Priority-Low | I find it hard to distinguish the active mission markers from the markers for potential missions that I could accept.
I understand you got rid of colours there for colourblind accessibility, but what about making them different shapes, or
use a different phase for the blinking? I don't mean frequency, too high might trigger epilepsy and too low makes
it hard to discern (there is still some usable range there, but that'd probably better be used for mission priorities),
but different phase at same frequency would work for proposed mission maker - of which there
is only one - vs. active markers - of which there are usually multiple. If the phase difference is noticably different
from pi, it will even be possible to tell which is which when both are in view with a bit of getting used to it.
I think something like PI/2 or 2/3 PI should do it (where time between two maxima of the same marker is 2*PI).
Logically, I'd think the phase of markers for missions already taken should lead the phase of markers for missions
that could potentially be taken. | 1.0 | mission markers are hard to distinguish - I find it hard to distinguish the active mission markers from the markers for potential missions that I could accept.
I understand you got rid of colours there for colourblind accessibility, but what about making them different shapes, or
use a different phase for the blinking? I don't mean frequency, too high might trigger epilepsy and too low makes
it hard to discern (there is still some usable range there, but that'd probably better be used for mission priorities),
but different phase at same frequency would work for proposed mission maker - of which there
is only one - vs. active markers - of which there are usually multiple. If the phase difference is noticably different
from pi, it will even be possible to tell which is which when both are in view with a bit of getting used to it.
I think something like PI/2 or 2/3 PI should do it (where time between two maxima of the same marker is 2*PI).
Logically, I'd think the phase of markers for missions already taken should lead the phase of markers for missions
that could potentially be taken. | non_defect | mission markers are hard to distinguish i find it hard to distinguish the active mission markers from the markers for potential missions that i could accept i understand you got rid of colours there for colourblind accessibility but what about making them different shapes or use a different phase for the blinking i don t mean frequency too high might trigger epilepsy and too low makes it hard to discern there is still some usable range there but that d probably better be used for mission priorities but different phase at same frequency would work for proposed mission maker of which there is only one vs active markers of which there are usually multiple if the phase difference is noticably different from pi it will even be possible to tell which is which when both are in view with a bit of getting used to it i think something like pi or pi should do it where time between two maxima of the same marker is pi logically i d think the phase of markers for missions already taken should lead the phase of markers for missions that could potentially be taken | 0 |
331,522 | 28,967,621,427 | IssuesEvent | 2023-05-10 08:59:38 | Joystream/joystream | https://api.github.com/repos/Joystream/joystream | opened | Fix direct channel payment flow | network-integration-test nara-network | Error is introduced in merging `nara` -> `crt_release`
output from processor:

failing flow: https://github.com/Joystream/joystream/actions/runs/4926868090/jobs/8817633132?pr=4749
| 1.0 | Fix direct channel payment flow - Error is introduced in merging `nara` -> `crt_release`
output from processor:

failing flow: https://github.com/Joystream/joystream/actions/runs/4926868090/jobs/8817633132?pr=4749
| non_defect | fix direct channel payment flow error is introduced in merging nara crt release output from processor failing flow | 0 |
50,900 | 13,187,954,763 | IssuesEvent | 2020-08-13 05:07:50 | icecube-trac/tix3 | https://api.github.com/repos/icecube-trac/tix3 | closed | [filterscripts] coordinate service -> astro (Trac #1641) | Migrated from Trac combo reconstruction defect | This import needs to change:
```text
File "/home/dschultz/Documents/combo/trunk/build_memory/lib/icecube/filterscripts/gcfilter.py", line 17, in GCFilter
from icecube import dataclasses, coordinate_service
ImportError: cannot import name coordinate_service
```
And this line in the file:
```text
add_gcfilter(tray,label=name,If=If,CorsikaMJDStart=coordinate_service.calendar_date_2_mjd(2013,07,01,00,00,00.),CorsikaMJDEnd=coordinate_service.calendar_date_2_mjd(2013,07,01,23,56,04.091)) ##One full MJD
```
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1641">https://code.icecube.wisc.edu/ticket/1641</a>, reported by david.schultz and owned by kjmeagher</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:57",
"description": "This import needs to change:\n{{{\n File \"/home/dschultz/Documents/combo/trunk/build_memory/lib/icecube/filterscripts/gcfilter.py\", line 17, in GCFilter\n from icecube import dataclasses, coordinate_service\nImportError: cannot import name coordinate_service\n}}}\n\nAnd this line in the file:\n{{{\nadd_gcfilter(tray,label=name,If=If,CorsikaMJDStart=coordinate_service.calendar_date_2_mjd(2013,07,01,00,00,00.),CorsikaMJDEnd=coordinate_service.calendar_date_2_mjd(2013,07,01,23,56,04.091)) ##One full MJD\n}}}",
"reporter": "david.schultz",
"cc": "",
"resolution": "fixed",
"_ts": "1550067117911749",
"component": "combo reconstruction",
"summary": "[filterscripts] coordinate service -> astro",
"priority": "blocker",
"keywords": "",
"time": "2016-04-17T17:21:55",
"milestone": "",
"owner": "kjmeagher",
"type": "defect"
}
```
</p>
</details>
| 1.0 | [filterscripts] coordinate service -> astro (Trac #1641) - This import needs to change:
```text
File "/home/dschultz/Documents/combo/trunk/build_memory/lib/icecube/filterscripts/gcfilter.py", line 17, in GCFilter
from icecube import dataclasses, coordinate_service
ImportError: cannot import name coordinate_service
```
And this line in the file:
```text
add_gcfilter(tray,label=name,If=If,CorsikaMJDStart=coordinate_service.calendar_date_2_mjd(2013,07,01,00,00,00.),CorsikaMJDEnd=coordinate_service.calendar_date_2_mjd(2013,07,01,23,56,04.091)) ##One full MJD
```
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1641">https://code.icecube.wisc.edu/ticket/1641</a>, reported by david.schultz and owned by kjmeagher</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:11:57",
"description": "This import needs to change:\n{{{\n File \"/home/dschultz/Documents/combo/trunk/build_memory/lib/icecube/filterscripts/gcfilter.py\", line 17, in GCFilter\n from icecube import dataclasses, coordinate_service\nImportError: cannot import name coordinate_service\n}}}\n\nAnd this line in the file:\n{{{\nadd_gcfilter(tray,label=name,If=If,CorsikaMJDStart=coordinate_service.calendar_date_2_mjd(2013,07,01,00,00,00.),CorsikaMJDEnd=coordinate_service.calendar_date_2_mjd(2013,07,01,23,56,04.091)) ##One full MJD\n}}}",
"reporter": "david.schultz",
"cc": "",
"resolution": "fixed",
"_ts": "1550067117911749",
"component": "combo reconstruction",
"summary": "[filterscripts] coordinate service -> astro",
"priority": "blocker",
"keywords": "",
"time": "2016-04-17T17:21:55",
"milestone": "",
"owner": "kjmeagher",
"type": "defect"
}
```
</p>
</details>
| defect | coordinate service astro trac this import needs to change text file home dschultz documents combo trunk build memory lib icecube filterscripts gcfilter py line in gcfilter from icecube import dataclasses coordinate service importerror cannot import name coordinate service and this line in the file text add gcfilter tray label name if if corsikamjdstart coordinate service calendar date mjd corsikamjdend coordinate service calendar date mjd one full mjd migrated from json status closed changetime description this import needs to change n n file home dschultz documents combo trunk build memory lib icecube filterscripts gcfilter py line in gcfilter n from icecube import dataclasses coordinate service nimporterror cannot import name coordinate service n n nand this line in the file n nadd gcfilter tray label name if if corsikamjdstart coordinate service calendar date mjd corsikamjdend coordinate service calendar date mjd one full mjd n reporter david schultz cc resolution fixed ts component combo reconstruction summary coordinate service astro priority blocker keywords time milestone owner kjmeagher type defect | 1 |
122,655 | 10,228,902,302 | IssuesEvent | 2019-08-17 07:28:33 | ballerina-platform/ballerina-lang | https://api.github.com/repos/ballerina-platform/ballerina-lang | closed | JDBC driver installation related info missing in JDBC api doc | Area/StandardLibs BetaTesting Type/Docs | **Description:**
Subject please in doc https://v1-0-0-alpha.ballerina.io/learn/api-docs/ballerina/java.jdbc/index.html
**Steps to reproduce:**
**Affected Versions:**
**OS, DB, other environment details and versions:**
**Related Issues (optional):**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
**Suggested Labels (optional):**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees (optional):**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
| 1.0 | JDBC driver installation related info missing in JDBC api doc - **Description:**
Subject please in doc https://v1-0-0-alpha.ballerina.io/learn/api-docs/ballerina/java.jdbc/index.html
**Steps to reproduce:**
**Affected Versions:**
**OS, DB, other environment details and versions:**
**Related Issues (optional):**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
**Suggested Labels (optional):**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees (optional):**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
| non_defect | jdbc driver installation related info missing in jdbc api doc description subject please in doc steps to reproduce affected versions os db other environment details and versions related issues optional suggested labels optional suggested assignees optional | 0 |
426,942 | 29,669,335,257 | IssuesEvent | 2023-06-11 07:51:16 | fedewf1/repositorio-tp2 | https://api.github.com/repos/fedewf1/repositorio-tp2 | closed | Issues 2 tp5 | documentation Diseño | **Para los archivos sucursales.html, contacto.html, sus derivados archivos html y los archivos java que empleen. Aplicar lo siguiente.
**No tener en cuenta los archivos nav, header y footer. Dado que estas son comunues a todos los archivos.
Considerar el uso de las clases bootstrap Containers, Filas, Columnas, Alineación, Formularios y componentes que considere necesario.
El sitio se debe adaptar (ser responsive) para los diferentes tamaños de dispositivos.
Adaptar las clases del paquete model para anotarlas con @Component.
En las clases del paquete controller realizar la inyeccion de dependencia de los objetos necesarios para el controlador.
Agregar documentacion interna al codigo de los archivos mencionados al principio.
Dentro del paquete model realice la agregación de las anotaciones necesarias para realizar la validación
de los atributos.
En los controladores que procesan las peticiones de formularios realice las adaptaciones necesarias para
capturar los errores.
Realizar las modificaciones dentro de los archivos html que representan formularios para poder mostrar
al usuario los mensajes referidos a las validaciones. | 1.0 | Issues 2 tp5 - **Para los archivos sucursales.html, contacto.html, sus derivados archivos html y los archivos java que empleen. Aplicar lo siguiente.
**No tener en cuenta los archivos nav, header y footer. Dado que estas son comunues a todos los archivos.
Considerar el uso de las clases bootstrap Containers, Filas, Columnas, Alineación, Formularios y componentes que considere necesario.
El sitio se debe adaptar (ser responsive) para los diferentes tamaños de dispositivos.
Adaptar las clases del paquete model para anotarlas con @Component.
En las clases del paquete controller realizar la inyeccion de dependencia de los objetos necesarios para el controlador.
Agregar documentacion interna al codigo de los archivos mencionados al principio.
Dentro del paquete model realice la agregación de las anotaciones necesarias para realizar la validación
de los atributos.
En los controladores que procesan las peticiones de formularios realice las adaptaciones necesarias para
capturar los errores.
Realizar las modificaciones dentro de los archivos html que representan formularios para poder mostrar
al usuario los mensajes referidos a las validaciones. | non_defect | issues para los archivos sucursales html contacto html sus derivados archivos html y los archivos java que empleen aplicar lo siguiente no tener en cuenta los archivos nav header y footer dado que estas son comunues a todos los archivos considerar el uso de las clases bootstrap containers filas columnas alineación formularios y componentes que considere necesario el sitio se debe adaptar ser responsive para los diferentes tamaños de dispositivos adaptar las clases del paquete model para anotarlas con component en las clases del paquete controller realizar la inyeccion de dependencia de los objetos necesarios para el controlador agregar documentacion interna al codigo de los archivos mencionados al principio dentro del paquete model realice la agregación de las anotaciones necesarias para realizar la validación de los atributos en los controladores que procesan las peticiones de formularios realice las adaptaciones necesarias para capturar los errores realizar las modificaciones dentro de los archivos html que representan formularios para poder mostrar al usuario los mensajes referidos a las validaciones | 0 |
52,093 | 13,211,387,926 | IssuesEvent | 2020-08-15 22:46:50 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | opened | [iceprod2] handle expiration of self-signed cert for webserver (Trac #1676) | Incomplete Migration Migrated from Trac defect iceprod | <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1676">https://code.icecube.wisc.edu/projects/icecube/ticket/1676</a>, reported by david.schultzand owned by david.schultz</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-05-09T21:55:16",
"_ts": "1462830916934929",
"description": "The default right now is to generate a self-signed cert with a duration of 30 days. We could either make this time infinitely large (several years), or deal with the problem of expiring certs.\n\nNote that just making a new cert isn't good enough, because already running jobs will still have the old cert.",
"reporter": "david.schultz",
"cc": "",
"resolution": "wontfix",
"time": "2016-04-29T16:41:30",
"component": "iceprod",
"summary": "[iceprod2] handle expiration of self-signed cert for webserver",
"priority": "major",
"keywords": "",
"milestone": "",
"owner": "david.schultz",
"type": "defect"
}
```
</p>
</details>
| 1.0 | [iceprod2] handle expiration of self-signed cert for webserver (Trac #1676) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1676">https://code.icecube.wisc.edu/projects/icecube/ticket/1676</a>, reported by david.schultzand owned by david.schultz</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-05-09T21:55:16",
"_ts": "1462830916934929",
"description": "The default right now is to generate a self-signed cert with a duration of 30 days. We could either make this time infinitely large (several years), or deal with the problem of expiring certs.\n\nNote that just making a new cert isn't good enough, because already running jobs will still have the old cert.",
"reporter": "david.schultz",
"cc": "",
"resolution": "wontfix",
"time": "2016-04-29T16:41:30",
"component": "iceprod",
"summary": "[iceprod2] handle expiration of self-signed cert for webserver",
"priority": "major",
"keywords": "",
"milestone": "",
"owner": "david.schultz",
"type": "defect"
}
```
</p>
</details>
| defect | handle expiration of self signed cert for webserver trac migrated from json status closed changetime ts description the default right now is to generate a self signed cert with a duration of days we could either make this time infinitely large several years or deal with the problem of expiring certs n nnote that just making a new cert isn t good enough because already running jobs will still have the old cert reporter david schultz cc resolution wontfix time component iceprod summary handle expiration of self signed cert for webserver priority major keywords milestone owner david schultz type defect | 1 |
30,869 | 6,335,280,954 | IssuesEvent | 2017-07-26 18:30:13 | scipy/scipy | https://api.github.com/repos/scipy/scipy | closed | optimize.BenchGlobal broken | Benchmarks defect | As reported on the mailing list, ``optimize.BenchGlobal`` benchmarks are broken currently:
```
python runtests.py --bench optimize.BenchGlobal --> fails on both my modified version (to add stochasticBB testing) and on the original scipy repository (errors described in the pastebin above)
python runtests.py --bench optimize.BenchLeastSquares --> works flawlessly
python runtests.py --bench optimize.BenchSmoothUnbounded --> works flawlessly
```
Traceback:
=============== =============== ==================== ============= ==================
-- result type / solver
--------------- ---------------------------------------------------------------------
test function success% / DE success% / basinh. <nfev> / DE <nfev> / basinh.
=============== =============== ==================== ============= ==================
AMGM failed failed failed failed
=============== =============== ==================== ============= ==================
[100.00%] ·····
For parameters: 'AMGM', 'success%', 'DE'
Traceback (most recent call last):
File "/Library/Frameworks/Python.framework/Versions/3.5/lib/python3.5/site-packages/asv/benchmark.py", line 773, in <module>
commands[mode](args)
File "/Library/Frameworks/Python.framework/Versions/3.5/lib/python3.5/site-packages/asv/benchmark.py", line 744, in main_run
skip = benchmark.do_setup()
File "/Library/Frameworks/Python.framework/Versions/3.5/lib/python3.5/site-packages/asv/benchmark.py", line 325, in do_setup
setup(*self._current_params)
TypeError: setup() takes 4 positional arguments but 5 were given
...
Version information:
Python 3.5
numpy 1.13.0
scipy 1.0.0.dev0+32d8e9e (current master)
asv 0.3.dev1123+32d09e76 (current master)
The problem seems to be that an extra parameter is passed from somewhere: ``setup() takes 4 positional arguments but 5 were given``. Checking those params shows that what's passed is ``name, ret_value, solver, None``. Unclear where the ``None`` is produced exactly. The ``BenchGlobal`` constructor which does not call ``super(BenchGlobal, self).__init__`` seems fragile.
Changing the signature of setup by appending a ``none`` parameter fixes things, but that doesn't look quite right. Last time this was modified is in gh-6400. @andyfaff what do you think?
| 1.0 | optimize.BenchGlobal broken - As reported on the mailing list, ``optimize.BenchGlobal`` benchmarks are broken currently:
```
python runtests.py --bench optimize.BenchGlobal --> fails on both my modified version (to add stochasticBB testing) and on the original scipy repository (errors described in the pastebin above)
python runtests.py --bench optimize.BenchLeastSquares --> works flawlessly
python runtests.py --bench optimize.BenchSmoothUnbounded --> works flawlessly
```
Traceback:
=============== =============== ==================== ============= ==================
-- result type / solver
--------------- ---------------------------------------------------------------------
test function success% / DE success% / basinh. <nfev> / DE <nfev> / basinh.
=============== =============== ==================== ============= ==================
AMGM failed failed failed failed
=============== =============== ==================== ============= ==================
[100.00%] ·····
For parameters: 'AMGM', 'success%', 'DE'
Traceback (most recent call last):
File "/Library/Frameworks/Python.framework/Versions/3.5/lib/python3.5/site-packages/asv/benchmark.py", line 773, in <module>
commands[mode](args)
File "/Library/Frameworks/Python.framework/Versions/3.5/lib/python3.5/site-packages/asv/benchmark.py", line 744, in main_run
skip = benchmark.do_setup()
File "/Library/Frameworks/Python.framework/Versions/3.5/lib/python3.5/site-packages/asv/benchmark.py", line 325, in do_setup
setup(*self._current_params)
TypeError: setup() takes 4 positional arguments but 5 were given
...
Version information:
Python 3.5
numpy 1.13.0
scipy 1.0.0.dev0+32d8e9e (current master)
asv 0.3.dev1123+32d09e76 (current master)
The problem seems to be that an extra parameter is passed from somewhere: ``setup() takes 4 positional arguments but 5 were given``. Checking those params shows that what's passed is ``name, ret_value, solver, None``. Unclear where the ``None`` is produced exactly. The ``BenchGlobal`` constructor which does not call ``super(BenchGlobal, self).__init__`` seems fragile.
Changing the signature of setup by appending a ``none`` parameter fixes things, but that doesn't look quite right. Last time this was modified is in gh-6400. @andyfaff what do you think?
| defect | optimize benchglobal broken as reported on the mailing list optimize benchglobal benchmarks are broken currently python runtests py bench optimize benchglobal fails on both my modified version to add stochasticbb testing and on the original scipy repository errors described in the pastebin above python runtests py bench optimize benchleastsquares works flawlessly python runtests py bench optimize benchsmoothunbounded works flawlessly traceback result type solver test function success de success basinh de basinh amgm failed failed failed failed ····· for parameters amgm success de traceback most recent call last file library frameworks python framework versions lib site packages asv benchmark py line in commands args file library frameworks python framework versions lib site packages asv benchmark py line in main run skip benchmark do setup file library frameworks python framework versions lib site packages asv benchmark py line in do setup setup self current params typeerror setup takes positional arguments but were given version information python numpy scipy current master asv current master the problem seems to be that an extra parameter is passed from somewhere setup takes positional arguments but were given checking those params shows that what s passed is name ret value solver none unclear where the none is produced exactly the benchglobal constructor which does not call super benchglobal self init seems fragile changing the signature of setup by appending a none parameter fixes things but that doesn t look quite right last time this was modified is in gh andyfaff what do you think | 1 |
175,254 | 27,815,663,231 | IssuesEvent | 2023-03-18 16:56:59 | Jade-ux/WomenTechConnect | https://api.github.com/repos/Jade-ux/WomenTechConnect | closed | Sign-up form - Design | design | Design and code separated into two cards but if you find it easier to go ahead and create it in the code straight away, that's great, feel free to assign both tickets to yourself
Wireframes for inspiration:

| 1.0 | Sign-up form - Design - Design and code separated into two cards but if you find it easier to go ahead and create it in the code straight away, that's great, feel free to assign both tickets to yourself
Wireframes for inspiration:

| non_defect | sign up form design design and code separated into two cards but if you find it easier to go ahead and create it in the code straight away that s great feel free to assign both tickets to yourself wireframes for inspiration | 0 |
250,590 | 18,895,221,407 | IssuesEvent | 2021-11-15 17:07:36 | apache/incubator-ponymail-foal | https://api.github.com/repos/apache/incubator-ponymail-foal | closed | What is this code trying to do? | bug documentation | https://github.com/apache/incubator-ponymail-foal/blob/6bbb4f49c99a22962afe7953b5fa06dfea341ec2/server/plugins/messages.py#L361-L364
AFAICT this will always result in the body being limited to BODY_MAXLEN + 1 (i.e. 201), as body_short is limited to SHORT_BODY_MAX_LEN+1 (i.e. 201) by the archiver.
So what is the point of the shorten option?
Why not just replace body with body_short? | 1.0 | What is this code trying to do? - https://github.com/apache/incubator-ponymail-foal/blob/6bbb4f49c99a22962afe7953b5fa06dfea341ec2/server/plugins/messages.py#L361-L364
AFAICT this will always result in the body being limited to BODY_MAXLEN + 1 (i.e. 201), as body_short is limited to SHORT_BODY_MAX_LEN+1 (i.e. 201) by the archiver.
So what is the point of the shorten option?
Why not just replace body with body_short? | non_defect | what is this code trying to do afaict this will always result in the body being limited to body maxlen i e as body short is limited to short body max len i e by the archiver so what is the point of the shorten option why not just replace body with body short | 0 |
65,819 | 19,707,685,609 | IssuesEvent | 2022-01-13 00:30:55 | jccastillo0007/eFacturaT | https://api.github.com/repos/jccastillo0007/eFacturaT | closed | CCP - misceláneos, no envía el número de registro ni residencia fiscal cuándo éstos se capturan, | defect | Estos campos se utilizan en traslados internacionales.
Aquí lo que se necesita es enviarlos al XML, cuando se capturen.
Por ejemplo, el conector como simplemente envía lo que incluye el archivo de texto, si lo hace:
<cartaporte20:Ubicacion DistanciaRecorrida="1000.0" FechaHoraSalidaLlegada="2021-05-10T10:00:00" NombreRemitenteDestinatario="PREMIER AUTOCOUNTRY SA DE CV" NumRegIdTrib="123456" RFCRemitenteDestinatario="XEXX010101000" ResidenciaFiscal="AFG" TipoUbicacion="Destino">
En el caso de la plataforma, no lo hace. Simplemente no lo envía, y el SAT pues marca error cuando el rfc del origen o destino es XEXX010101000, lo cual indica que debe acompañarse de estos 2 datos que no está incluyendo. | 1.0 | CCP - misceláneos, no envía el número de registro ni residencia fiscal cuándo éstos se capturan, - Estos campos se utilizan en traslados internacionales.
Aquí lo que se necesita es enviarlos al XML, cuando se capturen.
Por ejemplo, el conector como simplemente envía lo que incluye el archivo de texto, si lo hace:
<cartaporte20:Ubicacion DistanciaRecorrida="1000.0" FechaHoraSalidaLlegada="2021-05-10T10:00:00" NombreRemitenteDestinatario="PREMIER AUTOCOUNTRY SA DE CV" NumRegIdTrib="123456" RFCRemitenteDestinatario="XEXX010101000" ResidenciaFiscal="AFG" TipoUbicacion="Destino">
En el caso de la plataforma, no lo hace. Simplemente no lo envía, y el SAT pues marca error cuando el rfc del origen o destino es XEXX010101000, lo cual indica que debe acompañarse de estos 2 datos que no está incluyendo. | defect | ccp misceláneos no envía el número de registro ni residencia fiscal cuándo éstos se capturan estos campos se utilizan en traslados internacionales aquí lo que se necesita es enviarlos al xml cuando se capturen por ejemplo el conector como simplemente envía lo que incluye el archivo de texto si lo hace en el caso de la plataforma no lo hace simplemente no lo envía y el sat pues marca error cuando el rfc del origen o destino es lo cual indica que debe acompañarse de estos datos que no está incluyendo | 1 |
214,791 | 7,276,787,907 | IssuesEvent | 2018-02-21 17:21:31 | TylerConlee/slab | https://api.github.com/repos/TylerConlee/slab | closed | Add organization details to SLA ticket notifications | enhancement priority:normal | The organization that the ticket belongs to would be helpful for additional context on the ticket seen in the SLA notification. | 1.0 | Add organization details to SLA ticket notifications - The organization that the ticket belongs to would be helpful for additional context on the ticket seen in the SLA notification. | non_defect | add organization details to sla ticket notifications the organization that the ticket belongs to would be helpful for additional context on the ticket seen in the sla notification | 0 |
519,994 | 15,076,874,988 | IssuesEvent | 2021-02-05 05:47:10 | hassio-addons/addon-node-red | https://api.github.com/repos/hassio-addons/addon-node-red | closed | node-red-contrib-actionflows runtime errors (Deprecated Events) | priority-medium | # Problem/Motivation
The node-red component "node-red-contrib-actionflows" was written prior to the availability of "flows:started" and uses an older method of flow setup that is no longer supported in NodeRED. This triggers runtime deprecation notices.
## Expected behavior
No warnings, no indications on flow initiation.
## Actual behavior
Debug panel display on deployment AND deprecation notice on Node RED startup with the node in the configuration:
`[warn] [RED.events] Deprecated use of "nodes-started" event`
## Steps to reproduce
1. Add a flow with the node from "node-red-contrib-actionflows" in use (an inject -> node -> debug flow is fine for this purpose.
2. Display the Debug window in the side pane.
3. Deploy.
## Proposed changes
Adopt version 2.0.4 with fix.
PR to fix this deprecated event use has been merged by the maintainer.
PR here: https://github.com/Steveorevo/node-red-contrib-actionflows/pull/15
Main git here: https://github.com/Steveorevo/node-red-contrib-actionflows
| 1.0 | node-red-contrib-actionflows runtime errors (Deprecated Events) - # Problem/Motivation
The node-red component "node-red-contrib-actionflows" was written prior to the availability of "flows:started" and uses an older method of flow setup that is no longer supported in NodeRED. This triggers runtime deprecation notices.
## Expected behavior
No warnings, no indications on flow initiation.
## Actual behavior
Debug panel display on deployment AND deprecation notice on Node RED startup with the node in the configuration:
`[warn] [RED.events] Deprecated use of "nodes-started" event`
## Steps to reproduce
1. Add a flow with the node from "node-red-contrib-actionflows" in use (an inject -> node -> debug flow is fine for this purpose.
2. Display the Debug window in the side pane.
3. Deploy.
## Proposed changes
Adopt version 2.0.4 with fix.
PR to fix this deprecated event use has been merged by the maintainer.
PR here: https://github.com/Steveorevo/node-red-contrib-actionflows/pull/15
Main git here: https://github.com/Steveorevo/node-red-contrib-actionflows
| non_defect | node red contrib actionflows runtime errors deprecated events problem motivation the node red component node red contrib actionflows was written prior to the availability of flows started and uses an older method of flow setup that is no longer supported in nodered this triggers runtime deprecation notices expected behavior no warnings no indications on flow initiation actual behavior debug panel display on deployment and deprecation notice on node red startup with the node in the configuration deprecated use of nodes started event steps to reproduce add a flow with the node from node red contrib actionflows in use an inject node debug flow is fine for this purpose display the debug window in the side pane deploy proposed changes adopt version with fix pr to fix this deprecated event use has been merged by the maintainer pr here main git here | 0 |
48,935 | 7,466,453,561 | IssuesEvent | 2018-04-02 10:43:00 | scalameta/metals | https://api.github.com/repos/scalameta/metals | closed | Simplify installation | documentation installation | Currently, the installation steps are a sequence of several fairly fragile steps. https://github.com/scalameta/language-server/blob/master/BETA.md Any mistake could make nothing work, or half-work with no instructions on what's missing.
We should strive for as simple installation as possible but avoid 100% automagical setup that makes it difficult to track down what went wrong. For sbt, I propose
* addSbtPlugin
* document how to enable semanticdb-scalac compiler plugin
* `> metalsSetup` via sbt, downloads source jars, emit classpath and other stuff
TODOs:
* [ ] create sbt-metals plugin from project/MetalsPlugin.scala. We can depend on the plugin in the metals build by using `unmanagedSources in Compile ++= (baseDirectory.value.getParentFile / "sbt-metals").descedantsExcept("*.scala", ..).get`
| 1.0 | Simplify installation - Currently, the installation steps are a sequence of several fairly fragile steps. https://github.com/scalameta/language-server/blob/master/BETA.md Any mistake could make nothing work, or half-work with no instructions on what's missing.
We should strive for as simple installation as possible but avoid 100% automagical setup that makes it difficult to track down what went wrong. For sbt, I propose
* addSbtPlugin
* document how to enable semanticdb-scalac compiler plugin
* `> metalsSetup` via sbt, downloads source jars, emit classpath and other stuff
TODOs:
* [ ] create sbt-metals plugin from project/MetalsPlugin.scala. We can depend on the plugin in the metals build by using `unmanagedSources in Compile ++= (baseDirectory.value.getParentFile / "sbt-metals").descedantsExcept("*.scala", ..).get`
| non_defect | simplify installation currently the installation steps are a sequence of several fairly fragile steps any mistake could make nothing work or half work with no instructions on what s missing we should strive for as simple installation as possible but avoid automagical setup that makes it difficult to track down what went wrong for sbt i propose addsbtplugin document how to enable semanticdb scalac compiler plugin metalssetup via sbt downloads source jars emit classpath and other stuff todos create sbt metals plugin from project metalsplugin scala we can depend on the plugin in the metals build by using unmanagedsources in compile basedirectory value getparentfile sbt metals descedantsexcept scala get | 0 |
319,509 | 23,775,980,719 | IssuesEvent | 2022-09-01 20:59:51 | spacetelescope/drizzlepac | https://api.github.com/repos/spacetelescope/drizzlepac | closed | Need updates to DGEO warning message | Documentation | Hello! I'm working on a WFPC2 DrizzlePac ticket (INC0154128, for reference) and the user asked about the lengthy warning message below. I feel this needs to be updated for two reasons: 1) it provides users with an IRAF hedit command, and IRAF is deprecated, and 2) it should explain that for WFPC2 and other archival instruments, it is **fine** and **expected** to type 'c' and continue running drizzle. I don't believe that's clear at that moment. Thanks!!
>
> A 'DGEOFILE' keyword is present in the primary header but 'NPOLFILE' keyword was not found.
> This version of the software uses a new format for the residual distortion DGEO files.
> Please consult the instrument web pages for which reference files to download.
> A small (new style) dgeofile is needed ('_npl.fits' extension) and possibly a
> detector to image correction file ('_d2i.fits' extension).
> The names of these files must be added to the primary header either using the task XXXX
> or manually, for example:
>
> hedit u2pq0505t_c0m.fits[0] npolfile fname_npl.fits add+
> hedit u2pq0505t_c0m.fits[0] d2imfile fname_d2i.fits add+
>
> where fname_npl.fits is the name of the new style dgeo file and fname_d2i.fits is
> the name of the detector to image correction. After adding these keywords to the
> primary header, updatewcs must be run to update the science files:
>
> from stwcs import updatewcs
> updatewcs.updatewcs("u2pq0505t_c0m.fits")
>
> Alternatively you may choose to run astrodrizzle without DGEO and detector to image correction.
>
> To stop astrodrizzle and update the dgeo files, type 'q'.
> To continue running astrodrizzle without the non-polynomial distortion correction, type 'c': | 1.0 | Need updates to DGEO warning message - Hello! I'm working on a WFPC2 DrizzlePac ticket (INC0154128, for reference) and the user asked about the lengthy warning message below. I feel this needs to be updated for two reasons: 1) it provides users with an IRAF hedit command, and IRAF is deprecated, and 2) it should explain that for WFPC2 and other archival instruments, it is **fine** and **expected** to type 'c' and continue running drizzle. I don't believe that's clear at that moment. Thanks!!
>
> A 'DGEOFILE' keyword is present in the primary header but 'NPOLFILE' keyword was not found.
> This version of the software uses a new format for the residual distortion DGEO files.
> Please consult the instrument web pages for which reference files to download.
> A small (new style) dgeofile is needed ('_npl.fits' extension) and possibly a
> detector to image correction file ('_d2i.fits' extension).
> The names of these files must be added to the primary header either using the task XXXX
> or manually, for example:
>
> hedit u2pq0505t_c0m.fits[0] npolfile fname_npl.fits add+
> hedit u2pq0505t_c0m.fits[0] d2imfile fname_d2i.fits add+
>
> where fname_npl.fits is the name of the new style dgeo file and fname_d2i.fits is
> the name of the detector to image correction. After adding these keywords to the
> primary header, updatewcs must be run to update the science files:
>
> from stwcs import updatewcs
> updatewcs.updatewcs("u2pq0505t_c0m.fits")
>
> Alternatively you may choose to run astrodrizzle without DGEO and detector to image correction.
>
> To stop astrodrizzle and update the dgeo files, type 'q'.
> To continue running astrodrizzle without the non-polynomial distortion correction, type 'c': | non_defect | need updates to dgeo warning message hello i m working on a drizzlepac ticket for reference and the user asked about the lengthy warning message below i feel this needs to be updated for two reasons it provides users with an iraf hedit command and iraf is deprecated and it should explain that for and other archival instruments it is fine and expected to type c and continue running drizzle i don t believe that s clear at that moment thanks a dgeofile keyword is present in the primary header but npolfile keyword was not found this version of the software uses a new format for the residual distortion dgeo files please consult the instrument web pages for which reference files to download a small new style dgeofile is needed npl fits extension and possibly a detector to image correction file fits extension the names of these files must be added to the primary header either using the task xxxx or manually for example hedit fits npolfile fname npl fits add hedit fits fname fits add where fname npl fits is the name of the new style dgeo file and fname fits is the name of the detector to image correction after adding these keywords to the primary header updatewcs must be run to update the science files from stwcs import updatewcs updatewcs updatewcs fits alternatively you may choose to run astrodrizzle without dgeo and detector to image correction to stop astrodrizzle and update the dgeo files type q to continue running astrodrizzle without the non polynomial distortion correction type c | 0 |
103,135 | 11,340,124,514 | IssuesEvent | 2020-01-23 05:09:17 | wayexists02/hanyang-erica-robot-programming | https://api.github.com/repos/wayexists02/hanyang-erica-robot-programming | reopened | 코드론 실행법 | documentation | jylee 핫스팟(...)에 서버와 코드론을 연결한 후, 서버와 코드론에서 코드를 실행한다.
## 서버의 실행
코드론을 실행하기 앞서서 먼저 실행한다.
```roslaunch codrone_alpha launch.launch```
## 코드론의 실행
```roslaunch codrone_alpha_pi launch.launch``` | 1.0 | 코드론 실행법 - jylee 핫스팟(...)에 서버와 코드론을 연결한 후, 서버와 코드론에서 코드를 실행한다.
## 서버의 실행
코드론을 실행하기 앞서서 먼저 실행한다.
```roslaunch codrone_alpha launch.launch```
## 코드론의 실행
```roslaunch codrone_alpha_pi launch.launch``` | non_defect | 코드론 실행법 jylee 핫스팟 에 서버와 코드론을 연결한 후 서버와 코드론에서 코드를 실행한다 서버의 실행 코드론을 실행하기 앞서서 먼저 실행한다 roslaunch codrone alpha launch launch 코드론의 실행 roslaunch codrone alpha pi launch launch | 0 |
4,403 | 22,617,321,211 | IssuesEvent | 2022-06-30 00:20:29 | aws/aws-sam-cli | https://api.github.com/repos/aws/aws-sam-cli | closed | `sam sync` does not support custom bucket names | type/ux type/feature area/sam-config area/sync maintainer/need-followup area/accelerate |
### Description:
I don't use the default SAM bucket, I have my own. `sam sync` does not seem to support this.
### Steps to reproduce:
Do `sam init` and create the zip Python 3.9 "Hello World" template.
Create the following samconfig.toml
```toml
version = 0.1
[default]
[default.deploy]
[default.deploy.parameters]
stack_name = "sam-test"
s3_bucket = "mybucket"
s3_prefix = "sam-test"
region = "us-west-2"
capabilities = "CAPABILITY_IAM"
```
Run `sam build && sam deploy`, which succeeds.
### Observed result:
`sam sync --stack-name sam-test` gives the following output. You can see it's attempting to use the default managed SAM bucket.
```
2021-12-17 11:40:14,807 | Telemetry endpoint configured to be https://aws-serverless-tools-telemetry.us-west-2.amazonaws.com/metrics
2021-12-17 11:40:14,812 | Telemetry endpoint configured to be https://aws-serverless-tools-telemetry.us-west-2.amazonaws.com/metrics
2021-12-17 11:40:14,812 | Sending Telemetry: {'metrics': [{'templateWarning': {'requestId': '5e92f8cb-75e3-4793-81f8-faee808f01a7', 'installationId': '1ef32602-7319-4d1a-bc65-fb2419c3fe35', 'sessionId': 'eeb5b278-0298-446b-9bcc-43424c2cd44d', 'executionEnvironment': 'CLI', 'ci': False, 'pyversion': '3.8.12', 'samcliVersion': '1.36.0', 'awsProfileProvided': False, 'debugFlagProvided': True, 'region': '', 'warningName': 'CodeDeployWarning', 'warningCount': 0}}]}
2021-12-17 11:40:15,017 | Telemetry response: 200
2021-12-17 11:40:15,018 | Sending Telemetry: {'metrics': [{'templateWarning': {'requestId': 'd0f3bfd9-c6d7-40db-9c8b-337bf8efcd98', 'installationId': '1ef32602-7319-4d1a-bc65-fb2419c3fe35', 'sessionId': 'eeb5b278-0298-446b-9bcc-43424c2cd44d', 'executionEnvironment': 'CLI', 'ci': False, 'pyversion': '3.8.12', 'samcliVersion': '1.36.0', 'awsProfileProvided': False, 'debugFlagProvided': True, 'region': '', 'warningName': 'CodeDeployConditionWarning', 'warningCount': 0}}]}
2021-12-17 11:40:15,283 | Telemetry response: 200
2021-12-17 11:40:15,284 | Using config file: samconfig.toml, config environment: default
2021-12-17 11:40:15,284 | Expand command line arguments to:
2021-12-17 11:40:15,284 | --template_file=/Users/luhn/Code/audit/test/template.yaml --stack_name=sam-test --dependency_layer --capabilities=('CAPABILITY_NAMED_IAM', 'CAPABILITY_AUTO_EXPAND')
Managed S3 bucket: aws-sam-cli-managed-default-samclisourcebucket-1aupim17uw7m6
Default capabilities applied: ('CAPABILITY_NAMED_IAM', 'CAPABILITY_AUTO_EXPAND')
To override with customized capabilities, use --capabilities flag or set it in samconfig.toml
2021-12-17 11:40:16,112 | Using build directory as .aws-sam/auto-dependency-layer
2021-12-17 11:40:16,112 | Using build directory as .aws-sam/auto-dependency-layer
This feature is currently in beta. Visit the docs page to learn more about the AWS Beta terms https://aws.amazon.com/service-terms/.
The SAM CLI will use the AWS Lambda, Amazon API Gateway, and AWS StepFunctions APIs to upload your code without
performing a CloudFormation deployment. This will cause drift in your CloudFormation stack.
**The sync command should only be used against a development stack**.
Confirm that you are synchronizing a development stack and want to turn on beta features.
Enter Y to proceed with the command, or enter N to cancel:
[y/N]: 2021-12-17 11:40:17,467 | [33m
Experimental features are enabled for this session.
Visit the docs page to learn more about the AWS Beta terms https://aws.amazon.com/service-terms/.
[0m
2021-12-17 11:40:17,477 | No Parameters detected in the template
2021-12-17 11:40:17,499 | 2 stacks found in the template
2021-12-17 11:40:17,499 | No Parameters detected in the template
2021-12-17 11:40:17,510 | 2 resources found in the stack
2021-12-17 11:40:17,510 | No Parameters detected in the template
2021-12-17 11:40:17,519 | Found Serverless function with name='HelloWorldFunction' and CodeUri='hello_world/'
2021-12-17 11:40:17,519 | --base-dir is not presented, adjusting uri hello_world/ relative to /Users/luhn/Code/audit/test/template.yaml
2021-12-17 11:40:17,519 | No Parameters detected in the template
2021-12-17 11:40:17,538 | Executing the build using build context.
2021-12-17 11:40:17,538 | Instantiating build definitions
2021-12-17 11:40:17,540 | Same function build definition found, adding function (Previous: BuildDefinition(python3.9, /Users/luhn/Code/audit/test/hello_world, Zip, , d23e058e-cbff-4bce-85b2-09954cf33d29, {}, {}, x86_64, []), Current: BuildDefinition(python3.9, /Users/luhn/Code/audit/test/hello_world, Zip, , 85a07967-200c-4a31-81df-7700103e6ad7, {}, {}, x86_64, []), Function: Function(name='HelloWorldFunction', functionname='HelloWorldFunction', runtime='python3.9', memory=None, timeout=3, handler='app.lambda_handler', imageuri=None, packagetype='Zip', imageconfig=None, codeuri='/Users/luhn/Code/audit/test/hello_world', environment=None, rolearn=None, layers=[], events={'HelloWorld': {'Type': 'Api', 'Properties': {'Path': '/hello', 'Method': 'get', 'RestApiId': 'ServerlessRestApi'}}}, metadata=None, inlinecode=None, codesign_config_arn=None, architectures=['x86_64'], stack_path=''))
2021-12-17 11:40:17,541 | Async execution started
2021-12-17 11:40:17,541 | Invoking function functools.partial(<bound method CachedOrIncrementalBuildStrategyWrapper.build_single_function_definition of <samcli.lib.build.build_strategy.CachedOrIncrementalBuildStrategyWrapper object at 0x1056eb3d0>>, <samcli.lib.build.build_graph.FunctionBuildDefinition object at 0x1053468e0>)
2021-12-17 11:40:17,541 | Running incremental build for runtime python3.9 for build definition d23e058e-cbff-4bce-85b2-09954cf33d29
2021-12-17 11:40:17,541 | Waiting for async results
2021-12-17 11:40:17,541 | Manifest is not changed for d23e058e-cbff-4bce-85b2-09954cf33d29, running incremental build
2021-12-17 11:40:17,541 | Building codeuri: /Users/luhn/Code/audit/test/hello_world runtime: python3.9 metadata: {} architecture: x86_64 functions: ['HelloWorldFunction']
2021-12-17 11:40:17,541 | Building to following folder /Users/luhn/Code/audit/test/.aws-sam/auto-dependency-layer/HelloWorldFunction
2021-12-17 11:40:17,542 | Loading workflow module 'aws_lambda_builders.workflows'
2021-12-17 11:40:17,546 | Registering workflow 'PythonPipBuilder' with capability 'Capability(language='python', dependency_manager='pip', application_framework=None)'
2021-12-17 11:40:17,548 | Registering workflow 'NodejsNpmBuilder' with capability 'Capability(language='nodejs', dependency_manager='npm', application_framework=None)'
2021-12-17 11:40:17,549 | Registering workflow 'RubyBundlerBuilder' with capability 'Capability(language='ruby', dependency_manager='bundler', application_framework=None)'
2021-12-17 11:40:17,551 | Registering workflow 'GoDepBuilder' with capability 'Capability(language='go', dependency_manager='dep', application_framework=None)'
2021-12-17 11:40:17,553 | Registering workflow 'GoModulesBuilder' with capability 'Capability(language='go', dependency_manager='modules', application_framework=None)'
2021-12-17 11:40:17,555 | Registering workflow 'JavaGradleWorkflow' with capability 'Capability(language='java', dependency_manager='gradle', application_framework=None)'
2021-12-17 11:40:17,556 | Registering workflow 'JavaMavenWorkflow' with capability 'Capability(language='java', dependency_manager='maven', application_framework=None)'
2021-12-17 11:40:17,558 | Registering workflow 'DotnetCliPackageBuilder' with capability 'Capability(language='dotnet', dependency_manager='cli-package', application_framework=None)'
2021-12-17 11:40:17,559 | Registering workflow 'CustomMakeBuilder' with capability 'Capability(language='provided', dependency_manager=None, application_framework=None)'
2021-12-17 11:40:17,559 | Found workflow 'PythonPipBuilder' to support capabilities 'Capability(language='python', dependency_manager='pip', application_framework=None)'
2021-12-17 11:40:17,626 | Running workflow 'PythonPipBuilder'
2021-12-17 11:40:17,627 | Running PythonPipBuilder:CopySource
2021-12-17 11:40:17,629 | PythonPipBuilder:CopySource succeeded
2021-12-17 11:40:17,629 | Async execution completed
2021-12-17 11:40:17,630 | Auto creating dependency layer for each function resource into a nested stack
2021-12-17 11:40:17,630 | No Parameters detected in the template
2021-12-17 11:40:17,636 | 2 resources found in the stack sam-test
2021-12-17 11:40:17,636 | No Parameters detected in the template
2021-12-17 11:40:17,641 | Found Serverless function with name='HelloWorldFunction' and CodeUri='.aws-sam/auto-dependency-layer/HelloWorldFunction'
2021-12-17 11:40:17,641 | --base-dir is not presented, adjusting uri .aws-sam/auto-dependency-layer/HelloWorldFunction relative to /Users/luhn/Code/audit/test/template.yaml
Build Succeeded
Built Artifacts : .aws-sam/auto-dependency-layer
Built Template : .aws-sam/auto-dependency-layer/template.yaml
Commands you can use next
=========================
[*] Invoke Function: sam local invoke -t .aws-sam/auto-dependency-layer/template.yaml
[*] Test Function in the Cloud: sam sync --stack-name {stack-name} --watch
[*] Deploy: sam deploy --guided --template-file .aws-sam/auto-dependency-layer/template.yaml
2021-12-17 11:40:17,667 | Executing the packaging using package context.
2021-12-17 11:40:18,030 | Unable to export
Traceback (most recent call last):
File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/samcli/lib/package/s3_uploader.py", line 114, in upload
future.result()
File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/s3transfer/futures.py", line 106, in result
return self._coordinator.result()
File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/s3transfer/futures.py", line 265, in result
raise self._exception
File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/s3transfer/tasks.py", line 126, in __call__
return self._execute_main(kwargs)
File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/s3transfer/tasks.py", line 150, in _execute_main
return_value = self._main(**kwargs)
File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/s3transfer/upload.py", line 694, in _main
client.put_object(Bucket=bucket, Key=key, Body=body, **extra_args)
File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/botocore/client.py", line 391, in _api_call
return self._make_api_call(operation_name, kwargs)
File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/botocore/client.py", line 719, in _make_api_call
raise error_class(parsed_response, operation_name)
botocore.errorfactory.NoSuchBucket: An error occurred (NoSuchBucket) when calling the PutObject operation: The specified bucket does not exist
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/samcli/lib/package/packageable_resources.py", line 126, in export
self.do_export(resource_id, resource_dict, parent_dir)
File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/samcli/lib/package/packageable_resources.py", line 148, in do_export
uploaded_url = upload_local_artifacts(
File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/samcli/lib/package/utils.py", line 171, in upload_local_artifacts
return zip_and_upload(local_path, uploader, extension)
File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/samcli/lib/package/utils.py", line 189, in zip_and_upload
return uploader.upload_with_dedup(zip_file, precomputed_md5=md5_hash, extension=extension)
File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/samcli/lib/package/s3_uploader.py", line 143, in upload_with_dedup
return self.upload(file_name, remote_path)
File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/samcli/lib/package/s3_uploader.py", line 121, in upload
raise NoSuchBucketError(bucket_name=self.bucket_name) from ex
samcli.commands.package.exceptions.NoSuchBucketError:
S3 Bucket does not exist.
2021-12-17 11:40:18,033 | Sending Telemetry: {'metrics': [{'commandRunExperimental': {'requestId': '2898b15c-f378-4219-b192-da75e8d8e59d', 'installationId': '1ef32602-7319-4d1a-bc65-fb2419c3fe35', 'sessionId': 'eeb5b278-0298-446b-9bcc-43424c2cd44d', 'executionEnvironment': 'CLI', 'ci': False, 'pyversion': '3.8.12', 'samcliVersion': '1.36.0', 'awsProfileProvided': False, 'debugFlagProvided': True, 'region': '', 'commandName': 'sam sync', 'metricSpecificAttributes': {'experimentalAccelerate': True, 'experimentalAll': False}, 'duration': 3225, 'exitReason': 'ExportFailedError', 'exitCode': 1}}]}
2021-12-17 11:40:18,278 | Telemetry response: 200
Error: Unable to upload artifact HelloWorldFunction referenced by CodeUri parameter of HelloWorldFunction resource.
S3 Bucket does not exist.
```
### Expected result:
I would expect a) sync to honor the settings in samconfig.toml or b) a CLI flag to set the S3 bucket name.
### Additional environment details (Ex: Windows, Mac, Amazon Linux etc)
1. OS: Mac OS Monterey
2. If using SAM CLI, `sam --version`: `SAM CLI, version 1.36.0`
3. AWS region: us-west-2
| True | `sam sync` does not support custom bucket names -
### Description:
I don't use the default SAM bucket, I have my own. `sam sync` does not seem to support this.
### Steps to reproduce:
Do `sam init` and create the zip Python 3.9 "Hello World" template.
Create the following samconfig.toml
```toml
version = 0.1
[default]
[default.deploy]
[default.deploy.parameters]
stack_name = "sam-test"
s3_bucket = "mybucket"
s3_prefix = "sam-test"
region = "us-west-2"
capabilities = "CAPABILITY_IAM"
```
Run `sam build && sam deploy`, which succeeds.
### Observed result:
`sam sync --stack-name sam-test` gives the following output. You can see it's attempting to use the default managed SAM bucket.
```
2021-12-17 11:40:14,807 | Telemetry endpoint configured to be https://aws-serverless-tools-telemetry.us-west-2.amazonaws.com/metrics
2021-12-17 11:40:14,812 | Telemetry endpoint configured to be https://aws-serverless-tools-telemetry.us-west-2.amazonaws.com/metrics
2021-12-17 11:40:14,812 | Sending Telemetry: {'metrics': [{'templateWarning': {'requestId': '5e92f8cb-75e3-4793-81f8-faee808f01a7', 'installationId': '1ef32602-7319-4d1a-bc65-fb2419c3fe35', 'sessionId': 'eeb5b278-0298-446b-9bcc-43424c2cd44d', 'executionEnvironment': 'CLI', 'ci': False, 'pyversion': '3.8.12', 'samcliVersion': '1.36.0', 'awsProfileProvided': False, 'debugFlagProvided': True, 'region': '', 'warningName': 'CodeDeployWarning', 'warningCount': 0}}]}
2021-12-17 11:40:15,017 | Telemetry response: 200
2021-12-17 11:40:15,018 | Sending Telemetry: {'metrics': [{'templateWarning': {'requestId': 'd0f3bfd9-c6d7-40db-9c8b-337bf8efcd98', 'installationId': '1ef32602-7319-4d1a-bc65-fb2419c3fe35', 'sessionId': 'eeb5b278-0298-446b-9bcc-43424c2cd44d', 'executionEnvironment': 'CLI', 'ci': False, 'pyversion': '3.8.12', 'samcliVersion': '1.36.0', 'awsProfileProvided': False, 'debugFlagProvided': True, 'region': '', 'warningName': 'CodeDeployConditionWarning', 'warningCount': 0}}]}
2021-12-17 11:40:15,283 | Telemetry response: 200
2021-12-17 11:40:15,284 | Using config file: samconfig.toml, config environment: default
2021-12-17 11:40:15,284 | Expand command line arguments to:
2021-12-17 11:40:15,284 | --template_file=/Users/luhn/Code/audit/test/template.yaml --stack_name=sam-test --dependency_layer --capabilities=('CAPABILITY_NAMED_IAM', 'CAPABILITY_AUTO_EXPAND')
Managed S3 bucket: aws-sam-cli-managed-default-samclisourcebucket-1aupim17uw7m6
Default capabilities applied: ('CAPABILITY_NAMED_IAM', 'CAPABILITY_AUTO_EXPAND')
To override with customized capabilities, use --capabilities flag or set it in samconfig.toml
2021-12-17 11:40:16,112 | Using build directory as .aws-sam/auto-dependency-layer
2021-12-17 11:40:16,112 | Using build directory as .aws-sam/auto-dependency-layer
This feature is currently in beta. Visit the docs page to learn more about the AWS Beta terms https://aws.amazon.com/service-terms/.
The SAM CLI will use the AWS Lambda, Amazon API Gateway, and AWS StepFunctions APIs to upload your code without
performing a CloudFormation deployment. This will cause drift in your CloudFormation stack.
**The sync command should only be used against a development stack**.
Confirm that you are synchronizing a development stack and want to turn on beta features.
Enter Y to proceed with the command, or enter N to cancel:
[y/N]: 2021-12-17 11:40:17,467 | [33m
Experimental features are enabled for this session.
Visit the docs page to learn more about the AWS Beta terms https://aws.amazon.com/service-terms/.
[0m
2021-12-17 11:40:17,477 | No Parameters detected in the template
2021-12-17 11:40:17,499 | 2 stacks found in the template
2021-12-17 11:40:17,499 | No Parameters detected in the template
2021-12-17 11:40:17,510 | 2 resources found in the stack
2021-12-17 11:40:17,510 | No Parameters detected in the template
2021-12-17 11:40:17,519 | Found Serverless function with name='HelloWorldFunction' and CodeUri='hello_world/'
2021-12-17 11:40:17,519 | --base-dir is not presented, adjusting uri hello_world/ relative to /Users/luhn/Code/audit/test/template.yaml
2021-12-17 11:40:17,519 | No Parameters detected in the template
2021-12-17 11:40:17,538 | Executing the build using build context.
2021-12-17 11:40:17,538 | Instantiating build definitions
2021-12-17 11:40:17,540 | Same function build definition found, adding function (Previous: BuildDefinition(python3.9, /Users/luhn/Code/audit/test/hello_world, Zip, , d23e058e-cbff-4bce-85b2-09954cf33d29, {}, {}, x86_64, []), Current: BuildDefinition(python3.9, /Users/luhn/Code/audit/test/hello_world, Zip, , 85a07967-200c-4a31-81df-7700103e6ad7, {}, {}, x86_64, []), Function: Function(name='HelloWorldFunction', functionname='HelloWorldFunction', runtime='python3.9', memory=None, timeout=3, handler='app.lambda_handler', imageuri=None, packagetype='Zip', imageconfig=None, codeuri='/Users/luhn/Code/audit/test/hello_world', environment=None, rolearn=None, layers=[], events={'HelloWorld': {'Type': 'Api', 'Properties': {'Path': '/hello', 'Method': 'get', 'RestApiId': 'ServerlessRestApi'}}}, metadata=None, inlinecode=None, codesign_config_arn=None, architectures=['x86_64'], stack_path=''))
2021-12-17 11:40:17,541 | Async execution started
2021-12-17 11:40:17,541 | Invoking function functools.partial(<bound method CachedOrIncrementalBuildStrategyWrapper.build_single_function_definition of <samcli.lib.build.build_strategy.CachedOrIncrementalBuildStrategyWrapper object at 0x1056eb3d0>>, <samcli.lib.build.build_graph.FunctionBuildDefinition object at 0x1053468e0>)
2021-12-17 11:40:17,541 | Running incremental build for runtime python3.9 for build definition d23e058e-cbff-4bce-85b2-09954cf33d29
2021-12-17 11:40:17,541 | Waiting for async results
2021-12-17 11:40:17,541 | Manifest is not changed for d23e058e-cbff-4bce-85b2-09954cf33d29, running incremental build
2021-12-17 11:40:17,541 | Building codeuri: /Users/luhn/Code/audit/test/hello_world runtime: python3.9 metadata: {} architecture: x86_64 functions: ['HelloWorldFunction']
2021-12-17 11:40:17,541 | Building to following folder /Users/luhn/Code/audit/test/.aws-sam/auto-dependency-layer/HelloWorldFunction
2021-12-17 11:40:17,542 | Loading workflow module 'aws_lambda_builders.workflows'
2021-12-17 11:40:17,546 | Registering workflow 'PythonPipBuilder' with capability 'Capability(language='python', dependency_manager='pip', application_framework=None)'
2021-12-17 11:40:17,548 | Registering workflow 'NodejsNpmBuilder' with capability 'Capability(language='nodejs', dependency_manager='npm', application_framework=None)'
2021-12-17 11:40:17,549 | Registering workflow 'RubyBundlerBuilder' with capability 'Capability(language='ruby', dependency_manager='bundler', application_framework=None)'
2021-12-17 11:40:17,551 | Registering workflow 'GoDepBuilder' with capability 'Capability(language='go', dependency_manager='dep', application_framework=None)'
2021-12-17 11:40:17,553 | Registering workflow 'GoModulesBuilder' with capability 'Capability(language='go', dependency_manager='modules', application_framework=None)'
2021-12-17 11:40:17,555 | Registering workflow 'JavaGradleWorkflow' with capability 'Capability(language='java', dependency_manager='gradle', application_framework=None)'
2021-12-17 11:40:17,556 | Registering workflow 'JavaMavenWorkflow' with capability 'Capability(language='java', dependency_manager='maven', application_framework=None)'
2021-12-17 11:40:17,558 | Registering workflow 'DotnetCliPackageBuilder' with capability 'Capability(language='dotnet', dependency_manager='cli-package', application_framework=None)'
2021-12-17 11:40:17,559 | Registering workflow 'CustomMakeBuilder' with capability 'Capability(language='provided', dependency_manager=None, application_framework=None)'
2021-12-17 11:40:17,559 | Found workflow 'PythonPipBuilder' to support capabilities 'Capability(language='python', dependency_manager='pip', application_framework=None)'
2021-12-17 11:40:17,626 | Running workflow 'PythonPipBuilder'
2021-12-17 11:40:17,627 | Running PythonPipBuilder:CopySource
2021-12-17 11:40:17,629 | PythonPipBuilder:CopySource succeeded
2021-12-17 11:40:17,629 | Async execution completed
2021-12-17 11:40:17,630 | Auto creating dependency layer for each function resource into a nested stack
2021-12-17 11:40:17,630 | No Parameters detected in the template
2021-12-17 11:40:17,636 | 2 resources found in the stack sam-test
2021-12-17 11:40:17,636 | No Parameters detected in the template
2021-12-17 11:40:17,641 | Found Serverless function with name='HelloWorldFunction' and CodeUri='.aws-sam/auto-dependency-layer/HelloWorldFunction'
2021-12-17 11:40:17,641 | --base-dir is not presented, adjusting uri .aws-sam/auto-dependency-layer/HelloWorldFunction relative to /Users/luhn/Code/audit/test/template.yaml
Build Succeeded
Built Artifacts : .aws-sam/auto-dependency-layer
Built Template : .aws-sam/auto-dependency-layer/template.yaml
Commands you can use next
=========================
[*] Invoke Function: sam local invoke -t .aws-sam/auto-dependency-layer/template.yaml
[*] Test Function in the Cloud: sam sync --stack-name {stack-name} --watch
[*] Deploy: sam deploy --guided --template-file .aws-sam/auto-dependency-layer/template.yaml
2021-12-17 11:40:17,667 | Executing the packaging using package context.
2021-12-17 11:40:18,030 | Unable to export
Traceback (most recent call last):
File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/samcli/lib/package/s3_uploader.py", line 114, in upload
future.result()
File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/s3transfer/futures.py", line 106, in result
return self._coordinator.result()
File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/s3transfer/futures.py", line 265, in result
raise self._exception
File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/s3transfer/tasks.py", line 126, in __call__
return self._execute_main(kwargs)
File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/s3transfer/tasks.py", line 150, in _execute_main
return_value = self._main(**kwargs)
File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/s3transfer/upload.py", line 694, in _main
client.put_object(Bucket=bucket, Key=key, Body=body, **extra_args)
File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/botocore/client.py", line 391, in _api_call
return self._make_api_call(operation_name, kwargs)
File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/botocore/client.py", line 719, in _make_api_call
raise error_class(parsed_response, operation_name)
botocore.errorfactory.NoSuchBucket: An error occurred (NoSuchBucket) when calling the PutObject operation: The specified bucket does not exist
The above exception was the direct cause of the following exception:
Traceback (most recent call last):
File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/samcli/lib/package/packageable_resources.py", line 126, in export
self.do_export(resource_id, resource_dict, parent_dir)
File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/samcli/lib/package/packageable_resources.py", line 148, in do_export
uploaded_url = upload_local_artifacts(
File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/samcli/lib/package/utils.py", line 171, in upload_local_artifacts
return zip_and_upload(local_path, uploader, extension)
File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/samcli/lib/package/utils.py", line 189, in zip_and_upload
return uploader.upload_with_dedup(zip_file, precomputed_md5=md5_hash, extension=extension)
File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/samcli/lib/package/s3_uploader.py", line 143, in upload_with_dedup
return self.upload(file_name, remote_path)
File "/opt/homebrew/Cellar/aws-sam-cli/1.36.0/libexec/lib/python3.8/site-packages/samcli/lib/package/s3_uploader.py", line 121, in upload
raise NoSuchBucketError(bucket_name=self.bucket_name) from ex
samcli.commands.package.exceptions.NoSuchBucketError:
S3 Bucket does not exist.
2021-12-17 11:40:18,033 | Sending Telemetry: {'metrics': [{'commandRunExperimental': {'requestId': '2898b15c-f378-4219-b192-da75e8d8e59d', 'installationId': '1ef32602-7319-4d1a-bc65-fb2419c3fe35', 'sessionId': 'eeb5b278-0298-446b-9bcc-43424c2cd44d', 'executionEnvironment': 'CLI', 'ci': False, 'pyversion': '3.8.12', 'samcliVersion': '1.36.0', 'awsProfileProvided': False, 'debugFlagProvided': True, 'region': '', 'commandName': 'sam sync', 'metricSpecificAttributes': {'experimentalAccelerate': True, 'experimentalAll': False}, 'duration': 3225, 'exitReason': 'ExportFailedError', 'exitCode': 1}}]}
2021-12-17 11:40:18,278 | Telemetry response: 200
Error: Unable to upload artifact HelloWorldFunction referenced by CodeUri parameter of HelloWorldFunction resource.
S3 Bucket does not exist.
```
### Expected result:
I would expect a) sync to honor the settings in samconfig.toml or b) a CLI flag to set the S3 bucket name.
### Additional environment details (Ex: Windows, Mac, Amazon Linux etc)
1. OS: Mac OS Monterey
2. If using SAM CLI, `sam --version`: `SAM CLI, version 1.36.0`
3. AWS region: us-west-2
| non_defect | sam sync does not support custom bucket names description i don t use the default sam bucket i have my own sam sync does not seem to support this steps to reproduce do sam init and create the zip python hello world template create the following samconfig toml toml version stack name sam test bucket mybucket prefix sam test region us west capabilities capability iam run sam build sam deploy which succeeds observed result sam sync stack name sam test gives the following output you can see it s attempting to use the default managed sam bucket telemetry endpoint configured to be telemetry endpoint configured to be sending telemetry metrics telemetry response sending telemetry metrics telemetry response using config file samconfig toml config environment default expand command line arguments to template file users luhn code audit test template yaml stack name sam test dependency layer capabilities capability named iam capability auto expand managed bucket aws sam cli managed default samclisourcebucket default capabilities applied capability named iam capability auto expand to override with customized capabilities use capabilities flag or set it in samconfig toml using build directory as aws sam auto dependency layer using build directory as aws sam auto dependency layer this feature is currently in beta visit the docs page to learn more about the aws beta terms the sam cli will use the aws lambda amazon api gateway and aws stepfunctions apis to upload your code without performing a cloudformation deployment this will cause drift in your cloudformation stack the sync command should only be used against a development stack confirm that you are synchronizing a development stack and want to turn on beta features enter y to proceed with the command or enter n to cancel experimental features are enabled for this session visit the docs page to learn more about the aws beta terms no parameters detected in the template stacks found in the template no parameters detected in the template resources found in the stack no parameters detected in the template found serverless function with name helloworldfunction and codeuri hello world base dir is not presented adjusting uri hello world relative to users luhn code audit test template yaml no parameters detected in the template executing the build using build context instantiating build definitions same function build definition found adding function previous builddefinition users luhn code audit test hello world zip cbff current builddefinition users luhn code audit test hello world zip function function name helloworldfunction functionname helloworldfunction runtime memory none timeout handler app lambda handler imageuri none packagetype zip imageconfig none codeuri users luhn code audit test hello world environment none rolearn none layers events helloworld type api properties path hello method get restapiid serverlessrestapi metadata none inlinecode none codesign config arn none architectures stack path async execution started invoking function functools partial running incremental build for runtime for build definition cbff waiting for async results manifest is not changed for cbff running incremental build building codeuri users luhn code audit test hello world runtime metadata architecture functions building to following folder users luhn code audit test aws sam auto dependency layer helloworldfunction loading workflow module aws lambda builders workflows registering workflow pythonpipbuilder with capability capability language python dependency manager pip application framework none registering workflow nodejsnpmbuilder with capability capability language nodejs dependency manager npm application framework none registering workflow rubybundlerbuilder with capability capability language ruby dependency manager bundler application framework none registering workflow godepbuilder with capability capability language go dependency manager dep application framework none registering workflow gomodulesbuilder with capability capability language go dependency manager modules application framework none registering workflow javagradleworkflow with capability capability language java dependency manager gradle application framework none registering workflow javamavenworkflow with capability capability language java dependency manager maven application framework none registering workflow dotnetclipackagebuilder with capability capability language dotnet dependency manager cli package application framework none registering workflow custommakebuilder with capability capability language provided dependency manager none application framework none found workflow pythonpipbuilder to support capabilities capability language python dependency manager pip application framework none running workflow pythonpipbuilder running pythonpipbuilder copysource pythonpipbuilder copysource succeeded async execution completed auto creating dependency layer for each function resource into a nested stack no parameters detected in the template resources found in the stack sam test no parameters detected in the template found serverless function with name helloworldfunction and codeuri aws sam auto dependency layer helloworldfunction base dir is not presented adjusting uri aws sam auto dependency layer helloworldfunction relative to users luhn code audit test template yaml build succeeded built artifacts aws sam auto dependency layer built template aws sam auto dependency layer template yaml commands you can use next invoke function sam local invoke t aws sam auto dependency layer template yaml test function in the cloud sam sync stack name stack name watch deploy sam deploy guided template file aws sam auto dependency layer template yaml executing the packaging using package context unable to export traceback most recent call last file opt homebrew cellar aws sam cli libexec lib site packages samcli lib package uploader py line in upload future result file opt homebrew cellar aws sam cli libexec lib site packages futures py line in result return self coordinator result file opt homebrew cellar aws sam cli libexec lib site packages futures py line in result raise self exception file opt homebrew cellar aws sam cli libexec lib site packages tasks py line in call return self execute main kwargs file opt homebrew cellar aws sam cli libexec lib site packages tasks py line in execute main return value self main kwargs file opt homebrew cellar aws sam cli libexec lib site packages upload py line in main client put object bucket bucket key key body body extra args file opt homebrew cellar aws sam cli libexec lib site packages botocore client py line in api call return self make api call operation name kwargs file opt homebrew cellar aws sam cli libexec lib site packages botocore client py line in make api call raise error class parsed response operation name botocore errorfactory nosuchbucket an error occurred nosuchbucket when calling the putobject operation the specified bucket does not exist the above exception was the direct cause of the following exception traceback most recent call last file opt homebrew cellar aws sam cli libexec lib site packages samcli lib package packageable resources py line in export self do export resource id resource dict parent dir file opt homebrew cellar aws sam cli libexec lib site packages samcli lib package packageable resources py line in do export uploaded url upload local artifacts file opt homebrew cellar aws sam cli libexec lib site packages samcli lib package utils py line in upload local artifacts return zip and upload local path uploader extension file opt homebrew cellar aws sam cli libexec lib site packages samcli lib package utils py line in zip and upload return uploader upload with dedup zip file precomputed hash extension extension file opt homebrew cellar aws sam cli libexec lib site packages samcli lib package uploader py line in upload with dedup return self upload file name remote path file opt homebrew cellar aws sam cli libexec lib site packages samcli lib package uploader py line in upload raise nosuchbucketerror bucket name self bucket name from ex samcli commands package exceptions nosuchbucketerror bucket does not exist sending telemetry metrics telemetry response error unable to upload artifact helloworldfunction referenced by codeuri parameter of helloworldfunction resource bucket does not exist expected result i would expect a sync to honor the settings in samconfig toml or b a cli flag to set the bucket name additional environment details ex windows mac amazon linux etc os mac os monterey if using sam cli sam version sam cli version aws region us west | 0 |
8,043 | 2,611,449,702 | IssuesEvent | 2015-02-27 04:58:19 | chrsmith/hedgewars | https://api.github.com/repos/chrsmith/hedgewars | closed | Missing tombstones prevent game launch | auto-migrated Engine Priority-Medium Type-Defect | ```
(clone of
http://fireforge.net//tracker/?func=detail&atid=125&aid=274&group_id=11 )
Submitted by:
Christopher Vagnetoft (noccy80)
"Detailed description
Several times tonite a multiplayer game has failed to load since tombstones
haven't been found on one or more of the
players' computers. A missing tombstone should not cause the load to fail, but
rather revert to a default."
```
Original issue reported on code.google.com by `sheepyluva` on 28 Jun 2010 at 8:28 | 1.0 | Missing tombstones prevent game launch - ```
(clone of
http://fireforge.net//tracker/?func=detail&atid=125&aid=274&group_id=11 )
Submitted by:
Christopher Vagnetoft (noccy80)
"Detailed description
Several times tonite a multiplayer game has failed to load since tombstones
haven't been found on one or more of the
players' computers. A missing tombstone should not cause the load to fail, but
rather revert to a default."
```
Original issue reported on code.google.com by `sheepyluva` on 28 Jun 2010 at 8:28 | defect | missing tombstones prevent game launch clone of submitted by christopher vagnetoft detailed description several times tonite a multiplayer game has failed to load since tombstones haven t been found on one or more of the players computers a missing tombstone should not cause the load to fail but rather revert to a default original issue reported on code google com by sheepyluva on jun at | 1 |
43,514 | 23,270,872,739 | IssuesEvent | 2022-08-04 22:55:03 | quick-lint/quick-lint-js | https://api.github.com/repos/quick-lint/quick-lint-js | closed | Optimize Windows .exe icon | performance | dist/artwork/dusty-app.ico is 260 KiB. That's pretty big. Let's shrink it to bring the .exe file size down. | True | Optimize Windows .exe icon - dist/artwork/dusty-app.ico is 260 KiB. That's pretty big. Let's shrink it to bring the .exe file size down. | non_defect | optimize windows exe icon dist artwork dusty app ico is kib that s pretty big let s shrink it to bring the exe file size down | 0 |
99,881 | 8,714,077,373 | IssuesEvent | 2018-12-07 06:13:37 | actiontech/dble | https://api.github.com/repos/actiontech/dble | closed | System parameter `maxCon` is invalid after a connection failure | from auto_test resolve verified | * **dble version:**
5.6.29-dble-9.9.9.9-c53c3f7-20181116022818
* **preconditions :**
no
* **configs:**
**schema.xml**
```
<?xml version='1.0' encoding='utf-8'?>
<!DOCTYPE dble:schema SYSTEM "schema.dtd"><dble:schema xmlns:dble="http://dble.cloud/">
<schema dataNode="dn5" name="mytest" sqlMaxLimit="100">
<table dataNode="dn1,dn2,dn3,dn4" name="test" type="global" />
</schema>
<dataNode dataHost="172.100.9.5" database="db1" name="dn1" />
<dataNode dataHost="172.100.9.6" database="db1" name="dn2" />
<dataNode dataHost="172.100.9.5" database="db2" name="dn3" />
<dataNode dataHost="172.100.9.6" database="db2" name="dn4" />
<dataNode dataHost="172.100.9.5" database="db3" name="dn5" />
<dataHost balance="0" maxCon="1000" minCon="10" name="172.100.9.5" slaveThreshold="100" switchType="1">
<heartbeat>select user()</heartbeat>
<writeHost host="hostM1" password="111111" url="172.100.9.5:3306" user="test">
</writeHost>
</dataHost>
<dataHost balance="0" maxCon="1000" minCon="10" name="172.100.9.6" slaveThreshold="100" switchType="1">
<heartbeat>select user()</heartbeat>
<writeHost host="hostM2" password="111111" url="172.100.9.6:3306" user="test">
</writeHost>
</dataHost>
</dble:schema>
```
**rule.xml**
```
<?xml version='1.0' encoding='utf-8'?>
<!DOCTYPE dble:rule SYSTEM "rule.dtd"><dble:rule xmlns:dble="http://dble.cloud/">
<tableRule name="hash-two">
<rule>
<columns>id</columns>
<algorithm>two-long</algorithm>
</rule>
</tableRule>
<tableRule name="hash-three">
<rule>
<columns>id</columns>
<algorithm>three-long</algorithm>
</rule>
</tableRule>
<tableRule name="hash-four">
<rule>
<columns>id</columns>
<algorithm>four-long</algorithm>
</rule>
</tableRule>
<function class="Hash" name="two-long">
<property name="partitionCount">2</property>
<property name="partitionLength">1</property>
</function>
<function class="Hash" name="three-long">
<property name="partitionCount">3</property>
<property name="partitionLength">1</property>
</function>
<function class="Hash" name="four-long">
<property name="partitionCount">4</property>
<property name="partitionLength">1</property>
</function>
</dble:rule>
```
**server.xml**
```
<?xml version='1.0' encoding='utf-8'?>
<!DOCTYPE dble:server SYSTEM "server.dtd">
<dble:server xmlns:dble="http://dble.cloud/">
<system>
<property name="defaultSqlParser">druidparser</property>
<property name="useGlobleTableCheck">1</property>
<property name="processors">1</property>
<property name="processorExecutor">1</property>
<property name="maxCon">1</property>
</system>
<user name="root">
<property name="password">111111</property>
<property name="manager">true</property>
</user>
<user name="test">
<property name="password">test</property>
<property name="schemas">mytest</property>
<property name="maxCon">1</property>
</user>
<user name="action">
<property name="password">action</property>
<property name="schemas">mytest</property>
<property name="readOnly">true</property>
<property name="maxCon">1</property>
</user>
```
* **steps:**
step1. Do not close the connection after connecting to the client using the test user
```
[root@10-186-24-44 dble]# mysql -utest -ptest -h127.0.0.1 -P8066
mysql: [Warning] Using a password on the command line interface can be insecure.
Welcome to the MySQL monitor. Commands end with ; or \g.
Your MySQL connection id is 4
Server version: 5.6.29-dble-9.9.9.9-c53c3f7-20181116022818 dble Server (ActionTech)
Copyright (c) 2000, 2016, Oracle and/or its affiliates. All rights reserved.
Oracle is a registered trademark of Oracle Corporation and/or its
affiliates. Other names may be trademarks of their respective
owners.
Type 'help;' or '\h' for help. Type '\c' to clear the current input statement.
mysql>
```
step2. connect to the client using the test user
```
[root@10-186-24-44 dble]# mysql -utest -ptest -h127.0.0.1 -P8066
mysql: [Warning] Using a password on the command line interface can be insecure.
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
```
step3. connect to the client using the action user
```
[root@10-186-24-44 logs]# mysql -uaction -paction -h127.0.0.1 -P8066
mysql: [Warning] Using a password on the command line interface can be insecure.
Welcome to the MySQL monitor. Commands end with ; or \g.
Your MySQL connection id is 3
Server version: 5.6.29-dble-9.9.9.9-c53c3f7-20181116022818 dble Server (ActionTech)
Copyright (c) 2000, 2016, Oracle and/or its affiliates. All rights reserved.
Oracle is a registered trademark of Oracle Corporation and/or its
affiliates. Other names may be trademarks of their respective
owners.
Type 'help;' or '\h' for help. Type '\c' to clear the current input statement.
mysql>
```
* **expect result:**
1. step3: Connection failed
* **real result:**
1. step3: Connection successed
* **supplements:**
1.
/label ~BUG
| 1.0 | System parameter `maxCon` is invalid after a connection failure - * **dble version:**
5.6.29-dble-9.9.9.9-c53c3f7-20181116022818
* **preconditions :**
no
* **configs:**
**schema.xml**
```
<?xml version='1.0' encoding='utf-8'?>
<!DOCTYPE dble:schema SYSTEM "schema.dtd"><dble:schema xmlns:dble="http://dble.cloud/">
<schema dataNode="dn5" name="mytest" sqlMaxLimit="100">
<table dataNode="dn1,dn2,dn3,dn4" name="test" type="global" />
</schema>
<dataNode dataHost="172.100.9.5" database="db1" name="dn1" />
<dataNode dataHost="172.100.9.6" database="db1" name="dn2" />
<dataNode dataHost="172.100.9.5" database="db2" name="dn3" />
<dataNode dataHost="172.100.9.6" database="db2" name="dn4" />
<dataNode dataHost="172.100.9.5" database="db3" name="dn5" />
<dataHost balance="0" maxCon="1000" minCon="10" name="172.100.9.5" slaveThreshold="100" switchType="1">
<heartbeat>select user()</heartbeat>
<writeHost host="hostM1" password="111111" url="172.100.9.5:3306" user="test">
</writeHost>
</dataHost>
<dataHost balance="0" maxCon="1000" minCon="10" name="172.100.9.6" slaveThreshold="100" switchType="1">
<heartbeat>select user()</heartbeat>
<writeHost host="hostM2" password="111111" url="172.100.9.6:3306" user="test">
</writeHost>
</dataHost>
</dble:schema>
```
**rule.xml**
```
<?xml version='1.0' encoding='utf-8'?>
<!DOCTYPE dble:rule SYSTEM "rule.dtd"><dble:rule xmlns:dble="http://dble.cloud/">
<tableRule name="hash-two">
<rule>
<columns>id</columns>
<algorithm>two-long</algorithm>
</rule>
</tableRule>
<tableRule name="hash-three">
<rule>
<columns>id</columns>
<algorithm>three-long</algorithm>
</rule>
</tableRule>
<tableRule name="hash-four">
<rule>
<columns>id</columns>
<algorithm>four-long</algorithm>
</rule>
</tableRule>
<function class="Hash" name="two-long">
<property name="partitionCount">2</property>
<property name="partitionLength">1</property>
</function>
<function class="Hash" name="three-long">
<property name="partitionCount">3</property>
<property name="partitionLength">1</property>
</function>
<function class="Hash" name="four-long">
<property name="partitionCount">4</property>
<property name="partitionLength">1</property>
</function>
</dble:rule>
```
**server.xml**
```
<?xml version='1.0' encoding='utf-8'?>
<!DOCTYPE dble:server SYSTEM "server.dtd">
<dble:server xmlns:dble="http://dble.cloud/">
<system>
<property name="defaultSqlParser">druidparser</property>
<property name="useGlobleTableCheck">1</property>
<property name="processors">1</property>
<property name="processorExecutor">1</property>
<property name="maxCon">1</property>
</system>
<user name="root">
<property name="password">111111</property>
<property name="manager">true</property>
</user>
<user name="test">
<property name="password">test</property>
<property name="schemas">mytest</property>
<property name="maxCon">1</property>
</user>
<user name="action">
<property name="password">action</property>
<property name="schemas">mytest</property>
<property name="readOnly">true</property>
<property name="maxCon">1</property>
</user>
```
* **steps:**
step1. Do not close the connection after connecting to the client using the test user
```
[root@10-186-24-44 dble]# mysql -utest -ptest -h127.0.0.1 -P8066
mysql: [Warning] Using a password on the command line interface can be insecure.
Welcome to the MySQL monitor. Commands end with ; or \g.
Your MySQL connection id is 4
Server version: 5.6.29-dble-9.9.9.9-c53c3f7-20181116022818 dble Server (ActionTech)
Copyright (c) 2000, 2016, Oracle and/or its affiliates. All rights reserved.
Oracle is a registered trademark of Oracle Corporation and/or its
affiliates. Other names may be trademarks of their respective
owners.
Type 'help;' or '\h' for help. Type '\c' to clear the current input statement.
mysql>
```
step2. connect to the client using the test user
```
[root@10-186-24-44 dble]# mysql -utest -ptest -h127.0.0.1 -P8066
mysql: [Warning] Using a password on the command line interface can be insecure.
ERROR 2003 (HY000): Can't connect to MySQL server on '127.0.0.1' (111)
```
step3. connect to the client using the action user
```
[root@10-186-24-44 logs]# mysql -uaction -paction -h127.0.0.1 -P8066
mysql: [Warning] Using a password on the command line interface can be insecure.
Welcome to the MySQL monitor. Commands end with ; or \g.
Your MySQL connection id is 3
Server version: 5.6.29-dble-9.9.9.9-c53c3f7-20181116022818 dble Server (ActionTech)
Copyright (c) 2000, 2016, Oracle and/or its affiliates. All rights reserved.
Oracle is a registered trademark of Oracle Corporation and/or its
affiliates. Other names may be trademarks of their respective
owners.
Type 'help;' or '\h' for help. Type '\c' to clear the current input statement.
mysql>
```
* **expect result:**
1. step3: Connection failed
* **real result:**
1. step3: Connection successed
* **supplements:**
1.
/label ~BUG
| non_defect | system parameter maxcon is invalid after a connection failure dble version: dble preconditions : no configs: schema xml dble schema xmlns dble select user select user rule xml dble rule xmlns dble id two long id three long id four long server xml dble server xmlns dble druidparser true test mytest action mytest true steps: do not close the connection after connecting to the client using the test user mysql utest ptest mysql using a password on the command line interface can be insecure welcome to the mysql monitor commands end with or g your mysql connection id is server version dble dble server actiontech copyright c oracle and or its affiliates all rights reserved oracle is a registered trademark of oracle corporation and or its affiliates other names may be trademarks of their respective owners type help or h for help type c to clear the current input statement mysql connect to the client using the test user mysql utest ptest mysql using a password on the command line interface can be insecure error can t connect to mysql server on connect to the client using the action user mysql uaction paction mysql using a password on the command line interface can be insecure welcome to the mysql monitor commands end with or g your mysql connection id is server version dble dble server actiontech copyright c oracle and or its affiliates all rights reserved oracle is a registered trademark of oracle corporation and or its affiliates other names may be trademarks of their respective owners type help or h for help type c to clear the current input statement mysql expect result: connection failed real result: connection successed supplements: label bug | 0 |
16,097 | 2,871,807,576 | IssuesEvent | 2015-06-08 07:34:08 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | reopened | [TEST-FAILURE] MigrationAwareServiceTest.testPartitionDataSize_whenNodesStartedParallel_withSingleBackup | Team: Core Type: Defect | ```
java.lang.AssertionError: expected:<542> but was:<506>
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failNotEquals(Assert.java:834)
at org.junit.Assert.assertEquals(Assert.java:645)
```
https://hazelcast-l337.ci.cloudbees.com/job/Hazelcast-3.x-OpenJDK8-Quality-Outreach/com.hazelcast$hazelcast/72/testReport/junit/com.hazelcast.spi/MigrationAwareServiceTest/testPartitionDataSize_whenNodesStartedParallel_withSingleBackup/ | 1.0 | [TEST-FAILURE] MigrationAwareServiceTest.testPartitionDataSize_whenNodesStartedParallel_withSingleBackup - ```
java.lang.AssertionError: expected:<542> but was:<506>
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failNotEquals(Assert.java:834)
at org.junit.Assert.assertEquals(Assert.java:645)
```
https://hazelcast-l337.ci.cloudbees.com/job/Hazelcast-3.x-OpenJDK8-Quality-Outreach/com.hazelcast$hazelcast/72/testReport/junit/com.hazelcast.spi/MigrationAwareServiceTest/testPartitionDataSize_whenNodesStartedParallel_withSingleBackup/ | defect | migrationawareservicetest testpartitiondatasize whennodesstartedparallel withsinglebackup java lang assertionerror expected but was at org junit assert fail assert java at org junit assert failnotequals assert java at org junit assert assertequals assert java | 1 |
54,837 | 13,960,445,927 | IssuesEvent | 2020-10-24 21:07:12 | openzfs/zfs | https://api.github.com/repos/openzfs/zfs | opened | 'configure' fails on Kernel 5.8 with no module support | Status: Triage Needed Type: Defect | ### System information
Type | Version/Name
--- | ---
Distribution Name | Gentoo
Linux Kernel | 5.8+
Architecture | amd64
ZFS Version | 0.8.6-staging, 2.0.0-rc4
### Describe the problem you're observing
Configure fails when used in builtin mode for Linux 5.8 and 5.9 when those have no support for loadable modules.
It works on Linux 5.7 and lower with the same ZFS versions mentioned earlier.
The command I use:
```console
$ ./configure --with-linux=/path/to/linux-stable.git --enable-linux-builtin
[...]
checking kernel source directory... /path/to/linux-stable.git
checking kernel build directory... /path/to/linux-stable.git
checking kernel source version... 5.8.16
checking kernel file name for module symbols... NONE
checking whether modules can be built... no
configure: error:
*** Unable to build an empty module.
*** Please run 'make scripts' inside the kernel source tree.
```
### Include any warning/errors/backtraces from the system logs
Here is the extract of `config.log`:
```
configure:49874: checking whether modules can be built
configure:49935:
KBUILD_MODPOST_NOFINAL= KBUILD_MODPOST_WARN=
make modules -k -j8 -C /path/to/linux-stable.git
M=/path/to/zfs.git/build/conftest >build/conftest/build.log 2>&1
configure:49938: $? = 2
configure:49941: test -f build/conftest/conftest.o
configure:49944: $? = 1
configure:49952: result: no
configure:49959: error:
*** Unable to build an empty module.
*** Please run 'make scripts' inside the kernel source tree.
```
More interestingly, here is the content of `build/conftest/build.log`:
```
make: Entering directory '/path/to/linux-stable.git'
make[1]: *** No rule to make target '/path/to/zfs.git/build/conftest/modules.order', needed by '/path/to/zfs.git/build/conftest/Module.symvers'.
WARNING: Symbol version dump "Module.symvers" is missing.
Modules may not have dependencies or modversions.
make[1]: Target '__modpost' not remade because of errors.
make: *** [Makefile:1665: modules] Error 2
make: Leaving directory '/path/to/linux-stable.git'
```
### Further analysis in kernel changes
Of course this sudden breakage cannot be attributed to ZFS but this has a direct impact when you want to compile ZFS in a standalone kernel.
I did a little bisect in the kernel sources between 5.7 and 5.8 and found that the following commit is responsible:
- https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/commit/?id=436b2ac603d58504f38041a0cd8adb5aeace992b
It seems to be part of a bigger refactoring around modpost performed by Masahiro Yamada (@masahir0y).
There may be a workaround but honestly I don't fully understand what really causes this issue. | 1.0 | 'configure' fails on Kernel 5.8 with no module support - ### System information
Type | Version/Name
--- | ---
Distribution Name | Gentoo
Linux Kernel | 5.8+
Architecture | amd64
ZFS Version | 0.8.6-staging, 2.0.0-rc4
### Describe the problem you're observing
Configure fails when used in builtin mode for Linux 5.8 and 5.9 when those have no support for loadable modules.
It works on Linux 5.7 and lower with the same ZFS versions mentioned earlier.
The command I use:
```console
$ ./configure --with-linux=/path/to/linux-stable.git --enable-linux-builtin
[...]
checking kernel source directory... /path/to/linux-stable.git
checking kernel build directory... /path/to/linux-stable.git
checking kernel source version... 5.8.16
checking kernel file name for module symbols... NONE
checking whether modules can be built... no
configure: error:
*** Unable to build an empty module.
*** Please run 'make scripts' inside the kernel source tree.
```
### Include any warning/errors/backtraces from the system logs
Here is the extract of `config.log`:
```
configure:49874: checking whether modules can be built
configure:49935:
KBUILD_MODPOST_NOFINAL= KBUILD_MODPOST_WARN=
make modules -k -j8 -C /path/to/linux-stable.git
M=/path/to/zfs.git/build/conftest >build/conftest/build.log 2>&1
configure:49938: $? = 2
configure:49941: test -f build/conftest/conftest.o
configure:49944: $? = 1
configure:49952: result: no
configure:49959: error:
*** Unable to build an empty module.
*** Please run 'make scripts' inside the kernel source tree.
```
More interestingly, here is the content of `build/conftest/build.log`:
```
make: Entering directory '/path/to/linux-stable.git'
make[1]: *** No rule to make target '/path/to/zfs.git/build/conftest/modules.order', needed by '/path/to/zfs.git/build/conftest/Module.symvers'.
WARNING: Symbol version dump "Module.symvers" is missing.
Modules may not have dependencies or modversions.
make[1]: Target '__modpost' not remade because of errors.
make: *** [Makefile:1665: modules] Error 2
make: Leaving directory '/path/to/linux-stable.git'
```
### Further analysis in kernel changes
Of course this sudden breakage cannot be attributed to ZFS but this has a direct impact when you want to compile ZFS in a standalone kernel.
I did a little bisect in the kernel sources between 5.7 and 5.8 and found that the following commit is responsible:
- https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git/commit/?id=436b2ac603d58504f38041a0cd8adb5aeace992b
It seems to be part of a bigger refactoring around modpost performed by Masahiro Yamada (@masahir0y).
There may be a workaround but honestly I don't fully understand what really causes this issue. | defect | configure fails on kernel with no module support system information type version name distribution name gentoo linux kernel architecture zfs version staging describe the problem you re observing configure fails when used in builtin mode for linux and when those have no support for loadable modules it works on linux and lower with the same zfs versions mentioned earlier the command i use console configure with linux path to linux stable git enable linux builtin checking kernel source directory path to linux stable git checking kernel build directory path to linux stable git checking kernel source version checking kernel file name for module symbols none checking whether modules can be built no configure error unable to build an empty module please run make scripts inside the kernel source tree include any warning errors backtraces from the system logs here is the extract of config log configure checking whether modules can be built configure kbuild modpost nofinal kbuild modpost warn make modules k c path to linux stable git m path to zfs git build conftest build conftest build log configure configure test f build conftest conftest o configure configure result no configure error unable to build an empty module please run make scripts inside the kernel source tree more interestingly here is the content of build conftest build log make entering directory path to linux stable git make no rule to make target path to zfs git build conftest modules order needed by path to zfs git build conftest module symvers warning symbol version dump module symvers is missing modules may not have dependencies or modversions make target modpost not remade because of errors make error make leaving directory path to linux stable git further analysis in kernel changes of course this sudden breakage cannot be attributed to zfs but this has a direct impact when you want to compile zfs in a standalone kernel i did a little bisect in the kernel sources between and and found that the following commit is responsible it seems to be part of a bigger refactoring around modpost performed by masahiro yamada there may be a workaround but honestly i don t fully understand what really causes this issue | 1 |
18,541 | 2,615,173,145 | IssuesEvent | 2015-03-01 06:55:20 | chrsmith/html5rocks | https://api.github.com/repos/chrsmith/html5rocks | closed | typo in the fieldguide | auto-migrated fieldguide Milestone-Q12012 Priority-P3 Type-Bug | ```
"A web application providers a great user experience"
--------------------------^
at http://www.html5rocks.com/webappfieldguide/know-your-apps/site-vs-app/
```
Original issue reported on code.google.com by `mikenere...@gmail.com` on 15 Feb 2012 at 2:51 | 1.0 | typo in the fieldguide - ```
"A web application providers a great user experience"
--------------------------^
at http://www.html5rocks.com/webappfieldguide/know-your-apps/site-vs-app/
```
Original issue reported on code.google.com by `mikenere...@gmail.com` on 15 Feb 2012 at 2:51 | non_defect | typo in the fieldguide a web application providers a great user experience at original issue reported on code google com by mikenere gmail com on feb at | 0 |
217,939 | 16,891,531,463 | IssuesEvent | 2021-06-23 09:49:48 | hakehuang/zephyr | https://api.github.com/repos/hakehuang/zephyr | opened |
tests-ci :kernel.common.stack_protection_no_userspace.fatal : zephyr-v2.6.0-286-g46029914a7ac: mimxrt1015_evk: test Flash error
| area: Tests bug |
**Describe the bug**
kernel.common.stack_protection_no_userspace.fatal test is Flash error on zephyr-v2.6.0-286-g46029914a7ac on mimxrt1015_evk
see logs for details
**To Reproduce**
1.
```
scripts/twister --device-testing --device-serial /dev/ttyACM0 -p mimxrt1015_evk --testcase-root tests --sub-test kernel.common
```
2. See error
**Expected behavior**
test pass
**Impact**
**Logs and console output**
```
None
```
**Environment (please complete the following information):**
- OS: (e.g. Linux )
- Toolchain (e.g Zephyr SDK)
- Commit SHA or Version used: zephyr-v2.6.0-286-g46029914a7ac
| 1.0 |
tests-ci :kernel.common.stack_protection_no_userspace.fatal : zephyr-v2.6.0-286-g46029914a7ac: mimxrt1015_evk: test Flash error
-
**Describe the bug**
kernel.common.stack_protection_no_userspace.fatal test is Flash error on zephyr-v2.6.0-286-g46029914a7ac on mimxrt1015_evk
see logs for details
**To Reproduce**
1.
```
scripts/twister --device-testing --device-serial /dev/ttyACM0 -p mimxrt1015_evk --testcase-root tests --sub-test kernel.common
```
2. See error
**Expected behavior**
test pass
**Impact**
**Logs and console output**
```
None
```
**Environment (please complete the following information):**
- OS: (e.g. Linux )
- Toolchain (e.g Zephyr SDK)
- Commit SHA or Version used: zephyr-v2.6.0-286-g46029914a7ac
| non_defect | tests ci kernel common stack protection no userspace fatal zephyr evk test flash error describe the bug kernel common stack protection no userspace fatal test is flash error on zephyr on evk see logs for details to reproduce scripts twister device testing device serial dev p evk testcase root tests sub test kernel common see error expected behavior test pass impact logs and console output none environment please complete the following information os e g linux toolchain e g zephyr sdk commit sha or version used zephyr | 0 |
398,611 | 11,741,997,173 | IssuesEvent | 2020-03-11 23:16:32 | thaliawww/concrexit | https://api.github.com/repos/thaliawww/concrexit | closed | Toon dagnaam in events admin | events priority: low | In GitLab by njanssen on Mar 1, 2017, 20:05
Het zou fijn zijn als de naam van de dag (e.g. maandag, dinsdag, ...) in de datumweergaven van de backend staat.
Edit: Het gaat hierbij om de weergave van de evenementen. | 1.0 | Toon dagnaam in events admin - In GitLab by njanssen on Mar 1, 2017, 20:05
Het zou fijn zijn als de naam van de dag (e.g. maandag, dinsdag, ...) in de datumweergaven van de backend staat.
Edit: Het gaat hierbij om de weergave van de evenementen. | non_defect | toon dagnaam in events admin in gitlab by njanssen on mar het zou fijn zijn als de naam van de dag e g maandag dinsdag in de datumweergaven van de backend staat edit het gaat hierbij om de weergave van de evenementen | 0 |
2,245 | 2,712,129,537 | IssuesEvent | 2015-04-09 11:46:27 | HGustavs/LenaSYS | https://api.github.com/repos/HGustavs/LenaSYS | closed | The Back/Return button | CodeViewer | The Return/Back button in codeviewer isn't linked to a page, when clicked on.
<-a href="sectioned.php?courseid=UNK&coursevers=UNK"><-img src="../Shared/icons/Up.svg"-><-/a> The linked page is missing in the server(?) | 1.0 | The Back/Return button - The Return/Back button in codeviewer isn't linked to a page, when clicked on.
<-a href="sectioned.php?courseid=UNK&coursevers=UNK"><-img src="../Shared/icons/Up.svg"-><-/a> The linked page is missing in the server(?) | non_defect | the back return button the return back button in codeviewer isn t linked to a page when clicked on the linked page is missing in the server | 0 |
47,362 | 2,978,303,478 | IssuesEvent | 2015-07-16 04:47:27 | pombase/canto | https://api.github.com/repos/pombase/canto | closed | Do we need the New curs form? | admin low priority | Can we remove the New curs page (reached via the "Add ... Curation session" admin link)? It looks buggy in the test instance right now (and I ain't about to mess with what's in live), and I don't think it would be the most useful thing even if it got fixed. Doesn't seem like fun to generate hex-gibberish session IDs manually.
We can easily do without it, since we can find and add papers, and then push the button to create sessions for them, with autogenerated IDs creation dates. | 1.0 | Do we need the New curs form? - Can we remove the New curs page (reached via the "Add ... Curation session" admin link)? It looks buggy in the test instance right now (and I ain't about to mess with what's in live), and I don't think it would be the most useful thing even if it got fixed. Doesn't seem like fun to generate hex-gibberish session IDs manually.
We can easily do without it, since we can find and add papers, and then push the button to create sessions for them, with autogenerated IDs creation dates. | non_defect | do we need the new curs form can we remove the new curs page reached via the add curation session admin link it looks buggy in the test instance right now and i ain t about to mess with what s in live and i don t think it would be the most useful thing even if it got fixed doesn t seem like fun to generate hex gibberish session ids manually we can easily do without it since we can find and add papers and then push the button to create sessions for them with autogenerated ids creation dates | 0 |
32,386 | 6,767,416,567 | IssuesEvent | 2017-10-26 03:11:47 | Shopkeepers/Shopkeepers | https://api.github.com/repos/Shopkeepers/Shopkeepers | closed | With-Shop-creation near spawn->Witch blocked->Server restart->witch spawns but not editable | Defect fixed migrated | **Migrated from:** https://dev.bukkit.org/projects/shopkeepers/issues/111
**Originally posted by blablubbabc (Mar 29, 2013):**
I just discovered the following:- I have a zone in my world, where monster spawning is denied by a plugin.
- I tried setting up a witch shop there: it displayed me the message: "blabla, shop created, edit it!" but the witch of course was not spawned
- After a server restart (no reload), the witch is there.., BUT is not working ( I can't click it to open menu etc. .., but I still can't attack it..)
- I took a look at the save.yml and saw some other witch-shops also being there (which I must have tried to place them earlier in the same zone), BUT these witches are not placed at the locations saved, like the one described above..
- the witch, which was placed and can't be removed from ingame, is near the spawn (spawn chunk).. maybe this matters..Maybe you should block shop creation if an other plugin is blocking something or override the cancling of other plugins ( would be nice to be able to have witch-shops, while other monsters are blocked from spawning..THEN:
I just set debug to true in the config and reload the shopkeepers plugin via PluginManager (which seemed to work fine):
The witch is gone, and in the console there are messages for each not spawned witch:
"[WARNING] [Shopkeepers] Failed to spawn shopkeeper at world,434,71,722"Also:
Did you know that traders placed in tall grass are constantly ported back every 5 seconds for "being pushed", because their location was saved to the block above the tall grass and they fall through this block by their own all the time.. Maybe you could check for blocks like this (maybe add a list of "passable) block-types to the config.. peopel can then add water ect. to it if they want..) and set their location to the lowest possible location, going down all "passable" blocks on creation | 1.0 | With-Shop-creation near spawn->Witch blocked->Server restart->witch spawns but not editable - **Migrated from:** https://dev.bukkit.org/projects/shopkeepers/issues/111
**Originally posted by blablubbabc (Mar 29, 2013):**
I just discovered the following:- I have a zone in my world, where monster spawning is denied by a plugin.
- I tried setting up a witch shop there: it displayed me the message: "blabla, shop created, edit it!" but the witch of course was not spawned
- After a server restart (no reload), the witch is there.., BUT is not working ( I can't click it to open menu etc. .., but I still can't attack it..)
- I took a look at the save.yml and saw some other witch-shops also being there (which I must have tried to place them earlier in the same zone), BUT these witches are not placed at the locations saved, like the one described above..
- the witch, which was placed and can't be removed from ingame, is near the spawn (spawn chunk).. maybe this matters..Maybe you should block shop creation if an other plugin is blocking something or override the cancling of other plugins ( would be nice to be able to have witch-shops, while other monsters are blocked from spawning..THEN:
I just set debug to true in the config and reload the shopkeepers plugin via PluginManager (which seemed to work fine):
The witch is gone, and in the console there are messages for each not spawned witch:
"[WARNING] [Shopkeepers] Failed to spawn shopkeeper at world,434,71,722"Also:
Did you know that traders placed in tall grass are constantly ported back every 5 seconds for "being pushed", because their location was saved to the block above the tall grass and they fall through this block by their own all the time.. Maybe you could check for blocks like this (maybe add a list of "passable) block-types to the config.. peopel can then add water ect. to it if they want..) and set their location to the lowest possible location, going down all "passable" blocks on creation | defect | with shop creation near spawn witch blocked server restart witch spawns but not editable migrated from originally posted by blablubbabc mar i just discovered the following i have a zone in my world where monster spawning is denied by a plugin i tried setting up a witch shop there it displayed me the message blabla shop created edit it but the witch of course was not spawned after a server restart no reload the witch is there but is not working i can t click it to open menu etc but i still can t attack it i took a look at the save yml and saw some other witch shops also being there which i must have tried to place them earlier in the same zone but these witches are not placed at the locations saved like the one described above the witch which was placed and can t be removed from ingame is near the spawn spawn chunk maybe this matters maybe you should block shop creation if an other plugin is blocking something or override the cancling of other plugins would be nice to be able to have witch shops while other monsters are blocked from spawning then i just set debug to true in the config and reload the shopkeepers plugin via pluginmanager which seemed to work fine the witch is gone and in the console there are messages for each not spawned witch failed to spawn shopkeeper at world also did you know that traders placed in tall grass are constantly ported back every seconds for being pushed because their location was saved to the block above the tall grass and they fall through this block by their own all the time maybe you could check for blocks like this maybe add a list of passable block types to the config peopel can then add water ect to it if they want and set their location to the lowest possible location going down all passable blocks on creation | 1 |
45,111 | 11,589,787,925 | IssuesEvent | 2020-02-24 03:54:54 | GoogleContainerTools/skaffold | https://api.github.com/repos/GoogleContainerTools/skaffold | reopened | Get build timeout on local concurrent build | area/build kind/bug | ### Actual behavior
with `skaffold dev` i rarely get this sort of error:
```
Sending build context to Docker daemon 9.742MB
Step 1/10 : FROM node:10.15
FATA[1032] failed to build: build failed: building [us.gcr.io/replay-gaming/poker-api]: build artifact: unable to stream build output: Get https://registry-1.docker.io/v2/library/node/manifests/10.15: net/http: TLS handshake timeout
make[1]: *** [dependencies] Error 1
make: *** [init] Error 2
```
Keep in mind, that building of all images takes more than 15 minutes
### Information
- Skaffold version: 1.3.1
- Operating system: MacOs 10.15
- Contents of skaffold.yaml:
```yaml
apiVersion: skaffold/v2alpha3
kind: Config
build:
local:
concurrency: 0
tagPolicy:
sha256: {}
artifacts:
- image: us.gcr.io/replay-gaming/rails
docker:
dockerfile: docker/development/Dockerfile
context: rails
sync:
infer:
- 'app/**/*'
- 'bin/**/*'
- 'db/**/*'
- 'lib/**/*'
- 'public/**/*'
- 'script/**/*'
- 'spec/**/*'
- image: us.gcr.io/replay-gaming/poker-core
docker:
dockerfile: docker/development/Dockerfile
context: core
sync:
infer:
- 'bin/**/*'
- 'db/**/*'
- 'lib/**/*'
- 'system/**/*'
- 'schemas/**/*'
- 'spec/**/*'
- image: us.gcr.io/replay-gaming/poker-api
docker:
dockerfile: docker/development/Dockerfile
context: api
sync:
infer:
- 'bin/**/*'
- 'src/**/*'
- 'tests/**/*'
- image: us.gcr.io/replay-gaming/poker-client
docker:
dockerfile: docker/development/Dockerfile
context: client
sync:
infer:
- 'bin/**/*'
- 'public/**/*'
- 'src/**/*'
- image: us.gcr.io/replay-gaming/poker-scheduler
docker:
dockerfile: docker/development/Dockerfile
context: scheduler
sync:
infer:
- 'src/**/*'
- image: us.gcr.io/replay-gaming/go-eventsource
context: es
```
I assume that the root of the problem is in `concurrency` setting.
### Steps to reproduce the behavior
`skaffold dev`
This is flaky! | 1.0 | Get build timeout on local concurrent build - ### Actual behavior
with `skaffold dev` i rarely get this sort of error:
```
Sending build context to Docker daemon 9.742MB
Step 1/10 : FROM node:10.15
FATA[1032] failed to build: build failed: building [us.gcr.io/replay-gaming/poker-api]: build artifact: unable to stream build output: Get https://registry-1.docker.io/v2/library/node/manifests/10.15: net/http: TLS handshake timeout
make[1]: *** [dependencies] Error 1
make: *** [init] Error 2
```
Keep in mind, that building of all images takes more than 15 minutes
### Information
- Skaffold version: 1.3.1
- Operating system: MacOs 10.15
- Contents of skaffold.yaml:
```yaml
apiVersion: skaffold/v2alpha3
kind: Config
build:
local:
concurrency: 0
tagPolicy:
sha256: {}
artifacts:
- image: us.gcr.io/replay-gaming/rails
docker:
dockerfile: docker/development/Dockerfile
context: rails
sync:
infer:
- 'app/**/*'
- 'bin/**/*'
- 'db/**/*'
- 'lib/**/*'
- 'public/**/*'
- 'script/**/*'
- 'spec/**/*'
- image: us.gcr.io/replay-gaming/poker-core
docker:
dockerfile: docker/development/Dockerfile
context: core
sync:
infer:
- 'bin/**/*'
- 'db/**/*'
- 'lib/**/*'
- 'system/**/*'
- 'schemas/**/*'
- 'spec/**/*'
- image: us.gcr.io/replay-gaming/poker-api
docker:
dockerfile: docker/development/Dockerfile
context: api
sync:
infer:
- 'bin/**/*'
- 'src/**/*'
- 'tests/**/*'
- image: us.gcr.io/replay-gaming/poker-client
docker:
dockerfile: docker/development/Dockerfile
context: client
sync:
infer:
- 'bin/**/*'
- 'public/**/*'
- 'src/**/*'
- image: us.gcr.io/replay-gaming/poker-scheduler
docker:
dockerfile: docker/development/Dockerfile
context: scheduler
sync:
infer:
- 'src/**/*'
- image: us.gcr.io/replay-gaming/go-eventsource
context: es
```
I assume that the root of the problem is in `concurrency` setting.
### Steps to reproduce the behavior
`skaffold dev`
This is flaky! | non_defect | get build timeout on local concurrent build actual behavior with skaffold dev i rarely get this sort of error sending build context to docker daemon step from node fata failed to build build failed building build artifact unable to stream build output get net http tls handshake timeout make error make error keep in mind that building of all images takes more than minutes information skaffold version operating system macos contents of skaffold yaml yaml apiversion skaffold kind config build local concurrency tagpolicy artifacts image us gcr io replay gaming rails docker dockerfile docker development dockerfile context rails sync infer app bin db lib public script spec image us gcr io replay gaming poker core docker dockerfile docker development dockerfile context core sync infer bin db lib system schemas spec image us gcr io replay gaming poker api docker dockerfile docker development dockerfile context api sync infer bin src tests image us gcr io replay gaming poker client docker dockerfile docker development dockerfile context client sync infer bin public src image us gcr io replay gaming poker scheduler docker dockerfile docker development dockerfile context scheduler sync infer src image us gcr io replay gaming go eventsource context es i assume that the root of the problem is in concurrency setting steps to reproduce the behavior skaffold dev this is flaky | 0 |
98,444 | 29,870,547,119 | IssuesEvent | 2023-06-20 08:12:16 | GSS-Cogs/dd-cms | https://api.github.com/repos/GSS-Cogs/dd-cms | closed | Automate legend font change | chart builder high priority | At the moment the only way to update the legend font to the new font and size is to resave each chart. | 1.0 | Automate legend font change - At the moment the only way to update the legend font to the new font and size is to resave each chart. | non_defect | automate legend font change at the moment the only way to update the legend font to the new font and size is to resave each chart | 0 |
56,792 | 15,371,686,401 | IssuesEvent | 2021-03-02 10:18:47 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | com.hazelcast.json.MapPredicateJsonTest takes 20+ minutes to complete | Module: IMap Module: Query Source: Internal Team: Core Type: Defect Type: Test-Failure | Tests in the `com.hazelcast.json.MapPredicateJsonTest` take 20+ minutes to complete in the PR builder.
It might make sense to mark this as `SlowTest` but I feel like the tests there are pretty simple and should complete quickly.
When I run the test suite locally, I see that in some tests, I get the following logs regarding invocation retries and individual tests take 1+ min to complete. Sometimes, the same test takes 2-3 seconds to complete.
For example, for the `testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn`
```
17:13:40,188 INFO |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [LifecycleService] Thread-9 - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] [127.0.0.1]:5703 is STARTED
17:13:40,189 INFO |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [PartitionStateManager] hz.condescending_jennings.priority-generic-operation.thread-0 - [127.0.0.1]:5701 [dev] [4.2-SNAPSHOT] Initializing cluster partition table arrangement...
17:14:14,766 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=114, invocationTime=1611324854764 (2021-01-22 17:14:14.764), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=100, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService
17:14:15,280 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=124, invocationTime=1611324855279 (2021-01-22 17:14:15.279), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=110, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService
17:14:20,284 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=134, invocationTime=1611324860283 (2021-01-22 17:14:20.283), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=120, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService
17:14:25,288 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=144, invocationTime=1611324865286 (2021-01-22 17:14:25.286), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=130, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService
17:14:26,418 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=154, invocationTime=1611324866417 (2021-01-22 17:14:26.417), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=140, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService
17:14:30,806 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=164, invocationTime=1611324870805 (2021-01-22 17:14:30.805), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=150, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService
17:14:35,811 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=174, invocationTime=1611324875809 (2021-01-22 17:14:35.809), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=160, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService
17:14:37,846 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=184, invocationTime=1611324877845 (2021-01-22 17:14:37.845), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=170, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService
17:14:41,331 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=194, invocationTime=1611324881329 (2021-01-22 17:14:41.329), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=180, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService
```
It would be good to investigate the root cause of this. | 1.0 | com.hazelcast.json.MapPredicateJsonTest takes 20+ minutes to complete - Tests in the `com.hazelcast.json.MapPredicateJsonTest` take 20+ minutes to complete in the PR builder.
It might make sense to mark this as `SlowTest` but I feel like the tests there are pretty simple and should complete quickly.
When I run the test suite locally, I see that in some tests, I get the following logs regarding invocation retries and individual tests take 1+ min to complete. Sometimes, the same test takes 2-3 seconds to complete.
For example, for the `testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn`
```
17:13:40,188 INFO |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [LifecycleService] Thread-9 - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] [127.0.0.1]:5703 is STARTED
17:13:40,189 INFO |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [PartitionStateManager] hz.condescending_jennings.priority-generic-operation.thread-0 - [127.0.0.1]:5701 [dev] [4.2-SNAPSHOT] Initializing cluster partition table arrangement...
17:14:14,766 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=114, invocationTime=1611324854764 (2021-01-22 17:14:14.764), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=100, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService
17:14:15,280 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=124, invocationTime=1611324855279 (2021-01-22 17:14:15.279), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=110, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService
17:14:20,284 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=134, invocationTime=1611324860283 (2021-01-22 17:14:20.283), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=120, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService
17:14:25,288 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=144, invocationTime=1611324865286 (2021-01-22 17:14:25.286), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=130, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService
17:14:26,418 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=154, invocationTime=1611324866417 (2021-01-22 17:14:26.417), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=140, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService
17:14:30,806 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=164, invocationTime=1611324870805 (2021-01-22 17:14:30.805), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=150, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService
17:14:35,811 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=174, invocationTime=1611324875809 (2021-01-22 17:14:35.809), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=160, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService
17:14:37,846 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=184, invocationTime=1611324877845 (2021-01-22 17:14:37.845), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=170, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService
17:14:41,331 WARN |testNonTerminalAttributeIs_queriedWithEqualsNull_shouldNotReturn[inMemoryFormat: OBJECT, metadataPolicy: OFF]| - [Invocation] hz.vigorous_jennings.InvocationMonitorThread - [127.0.0.1]:5703 [dev] [4.2-SNAPSHOT] Retrying invocation: Invocation{op=com.hazelcast.map.impl.query.QueryOperation{serviceName='hz:impl:mapService', identityHash=1535734163, partitionId=-1, replicaIndex=0, callId=194, invocationTime=1611324881329 (2021-01-22 17:14:41.329), waitTimeout=-1, callTimeout=60000, tenantControl=com.hazelcast.spi.impl.tenantcontrol.NoopTenantControl@0, name=68ff64d9-4607-48eb-8bd3-6e0f53a872cb}, tryCount=250, tryPauseMillis=500, invokeCount=180, callTimeoutMillis=60000, firstInvocationTimeMs=1611324820203, firstInvocationTime='2021-01-22 17:13:40.203', lastHeartbeatMillis=0, lastHeartbeatTime='1970-01-01 02:00:00.000', target=null, pendingResponse={VOID}, backupsAcksExpected=-1, backupsAcksReceived=0, connection=null}, Reason: com.hazelcast.spi.exception.WrongTargetException: WrongTarget! local: Member [127.0.0.1]:5703 - 34c1e445-3684-452d-8ec4-bfa6476ecd26 this, expected-target: null, partitionId: -1, replicaIndex: 0, operation: com.hazelcast.map.impl.query.QueryOperation, service: hz:impl:mapService
```
It would be good to investigate the root cause of this. | defect | com hazelcast json mappredicatejsontest takes minutes to complete tests in the com hazelcast json mappredicatejsontest take minutes to complete in the pr builder it might make sense to mark this as slowtest but i feel like the tests there are pretty simple and should complete quickly when i run the test suite locally i see that in some tests i get the following logs regarding invocation retries and individual tests take min to complete sometimes the same test takes seconds to complete for example for the testnonterminalattributeis queriedwithequalsnull shouldnotreturn info testnonterminalattributeis queriedwithequalsnull shouldnotreturn thread is started info testnonterminalattributeis queriedwithequalsnull shouldnotreturn hz condescending jennings priority generic operation thread initializing cluster partition table arrangement warn testnonterminalattributeis queriedwithequalsnull shouldnotreturn hz vigorous jennings invocationmonitorthread retrying invocation invocation op com hazelcast map impl query queryoperation servicename hz impl mapservice identityhash partitionid replicaindex callid invocationtime waittimeout calltimeout tenantcontrol com hazelcast spi impl tenantcontrol nooptenantcontrol name trycount trypausemillis invokecount calltimeoutmillis firstinvocationtimems firstinvocationtime lastheartbeatmillis lastheartbeattime target null pendingresponse void backupsacksexpected backupsacksreceived connection null reason com hazelcast spi exception wrongtargetexception wrongtarget local member this expected target null partitionid replicaindex operation com hazelcast map impl query queryoperation service hz impl mapservice warn testnonterminalattributeis queriedwithequalsnull shouldnotreturn hz vigorous jennings invocationmonitorthread retrying invocation invocation op com hazelcast map impl query queryoperation servicename hz impl mapservice identityhash partitionid replicaindex callid invocationtime waittimeout calltimeout tenantcontrol com hazelcast spi impl tenantcontrol nooptenantcontrol name trycount trypausemillis invokecount calltimeoutmillis firstinvocationtimems firstinvocationtime lastheartbeatmillis lastheartbeattime target null pendingresponse void backupsacksexpected backupsacksreceived connection null reason com hazelcast spi exception wrongtargetexception wrongtarget local member this expected target null partitionid replicaindex operation com hazelcast map impl query queryoperation service hz impl mapservice warn testnonterminalattributeis queriedwithequalsnull shouldnotreturn hz vigorous jennings invocationmonitorthread retrying invocation invocation op com hazelcast map impl query queryoperation servicename hz impl mapservice identityhash partitionid replicaindex callid invocationtime waittimeout calltimeout tenantcontrol com hazelcast spi impl tenantcontrol nooptenantcontrol name trycount trypausemillis invokecount calltimeoutmillis firstinvocationtimems firstinvocationtime lastheartbeatmillis lastheartbeattime target null pendingresponse void backupsacksexpected backupsacksreceived connection null reason com hazelcast spi exception wrongtargetexception wrongtarget local member this expected target null partitionid replicaindex operation com hazelcast map impl query queryoperation service hz impl mapservice warn testnonterminalattributeis queriedwithequalsnull shouldnotreturn hz vigorous jennings invocationmonitorthread retrying invocation invocation op com hazelcast map impl query queryoperation servicename hz impl mapservice identityhash partitionid replicaindex callid invocationtime waittimeout calltimeout tenantcontrol com hazelcast spi impl tenantcontrol nooptenantcontrol name trycount trypausemillis invokecount calltimeoutmillis firstinvocationtimems firstinvocationtime lastheartbeatmillis lastheartbeattime target null pendingresponse void backupsacksexpected backupsacksreceived connection null reason com hazelcast spi exception wrongtargetexception wrongtarget local member this expected target null partitionid replicaindex operation com hazelcast map impl query queryoperation service hz impl mapservice warn testnonterminalattributeis queriedwithequalsnull shouldnotreturn hz vigorous jennings invocationmonitorthread retrying invocation invocation op com hazelcast map impl query queryoperation servicename hz impl mapservice identityhash partitionid replicaindex callid invocationtime waittimeout calltimeout tenantcontrol com hazelcast spi impl tenantcontrol nooptenantcontrol name trycount trypausemillis invokecount calltimeoutmillis firstinvocationtimems firstinvocationtime lastheartbeatmillis lastheartbeattime target null pendingresponse void backupsacksexpected backupsacksreceived connection null reason com hazelcast spi exception wrongtargetexception wrongtarget local member this expected target null partitionid replicaindex operation com hazelcast map impl query queryoperation service hz impl mapservice warn testnonterminalattributeis queriedwithequalsnull shouldnotreturn hz vigorous jennings invocationmonitorthread retrying invocation invocation op com hazelcast map impl query queryoperation servicename hz impl mapservice identityhash partitionid replicaindex callid invocationtime waittimeout calltimeout tenantcontrol com hazelcast spi impl tenantcontrol nooptenantcontrol name trycount trypausemillis invokecount calltimeoutmillis firstinvocationtimems firstinvocationtime lastheartbeatmillis lastheartbeattime target null pendingresponse void backupsacksexpected backupsacksreceived connection null reason com hazelcast spi exception wrongtargetexception wrongtarget local member this expected target null partitionid replicaindex operation com hazelcast map impl query queryoperation service hz impl mapservice warn testnonterminalattributeis queriedwithequalsnull shouldnotreturn hz vigorous jennings invocationmonitorthread retrying invocation invocation op com hazelcast map impl query queryoperation servicename hz impl mapservice identityhash partitionid replicaindex callid invocationtime waittimeout calltimeout tenantcontrol com hazelcast spi impl tenantcontrol nooptenantcontrol name trycount trypausemillis invokecount calltimeoutmillis firstinvocationtimems firstinvocationtime lastheartbeatmillis lastheartbeattime target null pendingresponse void backupsacksexpected backupsacksreceived connection null reason com hazelcast spi exception wrongtargetexception wrongtarget local member this expected target null partitionid replicaindex operation com hazelcast map impl query queryoperation service hz impl mapservice warn testnonterminalattributeis queriedwithequalsnull shouldnotreturn hz vigorous jennings invocationmonitorthread retrying invocation invocation op com hazelcast map impl query queryoperation servicename hz impl mapservice identityhash partitionid replicaindex callid invocationtime waittimeout calltimeout tenantcontrol com hazelcast spi impl tenantcontrol nooptenantcontrol name trycount trypausemillis invokecount calltimeoutmillis firstinvocationtimems firstinvocationtime lastheartbeatmillis lastheartbeattime target null pendingresponse void backupsacksexpected backupsacksreceived connection null reason com hazelcast spi exception wrongtargetexception wrongtarget local member this expected target null partitionid replicaindex operation com hazelcast map impl query queryoperation service hz impl mapservice warn testnonterminalattributeis queriedwithequalsnull shouldnotreturn hz vigorous jennings invocationmonitorthread retrying invocation invocation op com hazelcast map impl query queryoperation servicename hz impl mapservice identityhash partitionid replicaindex callid invocationtime waittimeout calltimeout tenantcontrol com hazelcast spi impl tenantcontrol nooptenantcontrol name trycount trypausemillis invokecount calltimeoutmillis firstinvocationtimems firstinvocationtime lastheartbeatmillis lastheartbeattime target null pendingresponse void backupsacksexpected backupsacksreceived connection null reason com hazelcast spi exception wrongtargetexception wrongtarget local member this expected target null partitionid replicaindex operation com hazelcast map impl query queryoperation service hz impl mapservice it would be good to investigate the root cause of this | 1 |
481,190 | 13,881,750,872 | IssuesEvent | 2020-10-18 02:23:50 | apcountryman/avr-libcpp | https://api.github.com/repos/apcountryman/avr-libcpp | closed | Add util/delay | priority-normal status-in_revision type-feature | Add `util/delay` header (`system/util/delay`), and associated header/source implementation files (`include/util/delay.h`, and `source/util/delay.cc`). | 1.0 | Add util/delay - Add `util/delay` header (`system/util/delay`), and associated header/source implementation files (`include/util/delay.h`, and `source/util/delay.cc`). | non_defect | add util delay add util delay header system util delay and associated header source implementation files include util delay h and source util delay cc | 0 |
21,573 | 3,520,005,485 | IssuesEvent | 2016-01-12 19:03:00 | WilliamOckham/hunpos | https://api.github.com/repos/WilliamOckham/hunpos | closed | tagging erroneous in some cases | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
Tag a sample corpus
What is the expected output? What do you see instead?
I wrote in for example:
got expected
I PRP
do VBP
work NN <-- VB
. SENT
What version of the product are you using? On what operating system?
1.2.8, linux mint
Please provide any additional information below.
```
Original issue reported on code.google.com by `krie...@gmx.de` on 19 Dec 2010 at 8:59 | 1.0 | tagging erroneous in some cases - ```
What steps will reproduce the problem?
Tag a sample corpus
What is the expected output? What do you see instead?
I wrote in for example:
got expected
I PRP
do VBP
work NN <-- VB
. SENT
What version of the product are you using? On what operating system?
1.2.8, linux mint
Please provide any additional information below.
```
Original issue reported on code.google.com by `krie...@gmx.de` on 19 Dec 2010 at 8:59 | defect | tagging erroneous in some cases what steps will reproduce the problem tag a sample corpus what is the expected output what do you see instead i wrote in for example got expected i prp do vbp work nn vb sent what version of the product are you using on what operating system linux mint please provide any additional information below original issue reported on code google com by krie gmx de on dec at | 1 |
486,981 | 14,017,401,933 | IssuesEvent | 2020-10-29 15:36:47 | GQCG/GQCP | https://api.github.com/repos/GQCG/GQCP | opened | Enable the evaluation of `GSQOperators` in a frozen-core spin-unresolved ONV basis | C++ complexity: intermediate priority: low theory | In recent refactors (#688), we had to temporarily disable some of the CI functionality. This issue tracks the re-enabling of the API to evaluate restricted operators in a seniority-zero ONV basis. | 1.0 | Enable the evaluation of `GSQOperators` in a frozen-core spin-unresolved ONV basis - In recent refactors (#688), we had to temporarily disable some of the CI functionality. This issue tracks the re-enabling of the API to evaluate restricted operators in a seniority-zero ONV basis. | non_defect | enable the evaluation of gsqoperators in a frozen core spin unresolved onv basis in recent refactors we had to temporarily disable some of the ci functionality this issue tracks the re enabling of the api to evaluate restricted operators in a seniority zero onv basis | 0 |
28,303 | 5,239,168,291 | IssuesEvent | 2017-01-31 08:53:26 | pexcn/tb-tun | https://api.github.com/repos/pexcn/tb-tun | closed | MTU and ICMP issue | auto-migrated Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1.
ICMP package on IPv4 should be reflected to the IPv6 according to RFC2893.
But TB-TUN has nothing to do with ICMP.
2.
According to RFC3506, "If the IPv6 MTU size proves to be too large for some
intermediate IPv4 subnet, IPv4 fragmentation will ensue....The IPv4 'do not
fragment' bit *SHOULD NOT* be set in the encapsulating IPv4 header." But
TB-TUN disobey these 2 rules. It causes IPv6 packages missing data and did
not report any error by ICMP when IPv6 MTU is to large.
What is the expected output? What do you see instead?
Please use labels and text to provide additional information.
```
Original issue reported on code.google.com by `lostrive...@gmail.com` on 25 Apr 2010 at 5:50
| 1.0 | MTU and ICMP issue - ```
What steps will reproduce the problem?
1.
ICMP package on IPv4 should be reflected to the IPv6 according to RFC2893.
But TB-TUN has nothing to do with ICMP.
2.
According to RFC3506, "If the IPv6 MTU size proves to be too large for some
intermediate IPv4 subnet, IPv4 fragmentation will ensue....The IPv4 'do not
fragment' bit *SHOULD NOT* be set in the encapsulating IPv4 header." But
TB-TUN disobey these 2 rules. It causes IPv6 packages missing data and did
not report any error by ICMP when IPv6 MTU is to large.
What is the expected output? What do you see instead?
Please use labels and text to provide additional information.
```
Original issue reported on code.google.com by `lostrive...@gmail.com` on 25 Apr 2010 at 5:50
| defect | mtu and icmp issue what steps will reproduce the problem icmp package on should be reflected to the according to but tb tun has nothing to do with icmp according to if the mtu size proves to be too large for some intermediate subnet fragmentation will ensue the do not fragment bit should not be set in the encapsulating header but tb tun disobey these rules it causes packages missing data and did not report any error by icmp when mtu is to large what is the expected output what do you see instead please use labels and text to provide additional information original issue reported on code google com by lostrive gmail com on apr at | 1 |
3,866 | 2,610,083,167 | IssuesEvent | 2015-02-26 18:25:25 | chrsmith/dsdsdaadf | https://api.github.com/repos/chrsmith/dsdsdaadf | opened | 祛除青春痘深圳 | auto-migrated Priority-Medium Type-Defect | ```
祛除青春痘深圳【深圳韩方科颜全国热线400-869-1818,24小时QQ4
008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘方��
�—韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方科�
��专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健康
祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业治��
�粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘痘�
��
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 6:47 | 1.0 | 祛除青春痘深圳 - ```
祛除青春痘深圳【深圳韩方科颜全国热线400-869-1818,24小时QQ4
008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘方��
�—韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方科�
��专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健康
祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业治��
�粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘痘�
��
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 6:47 | defect | 祛除青春痘深圳 祛除青春痘深圳【 , 】深圳韩方科颜专业祛痘连锁机构,机构以韩国秘方�� �—韩方科颜这一国妆准字号治疗型权威,祛痘佳品,韩方科� ��专业祛痘连锁机构,采用韩国秘方配合专业“不反弹”健康 祛痘技术并结合先进“先进豪华彩光”仪,开创国内专业治�� �粉刺、痤疮签约包治先河,成功消除了许多顾客脸上的痘痘� �� original issue reported on code google com by szft com on may at | 1 |
59,514 | 24,802,375,927 | IssuesEvent | 2022-10-24 23:18:13 | Azure/azure-sdk-for-python | https://api.github.com/repos/Azure/azure-sdk-for-python | closed | ApplicationInsights does not work out of the box when using ClientSecretCredentials | feature-request Monitor Service Attention Client Monitor - ApplicationInsights | - **Package Name**: azure-applicationinsights
- **Package Version**: 0.1.0
- **Operating System**: Linux
- **Python Version**: 3.8.7
**Describe the bug**
The applicationInsights client requires:
- the identity credential must have a signed_session (that is missing from the azure.identity.ClientSecretCredential) so a [wrapper](https://gist.github.com/lmazuel/cc683d82ea1d7b40208de7c9fc8de59d) has to be used to make it work.
- AppId and AppKey are needed to fetch queries succesfully: To pass the app key, the header needs to be modified in the applicationInsightDataClient
``` python
credential = ClientSecretCredential(
client_id=os.environ.get(env.ENV_APP_CLIENT_ID),
client_secret=os.environ.get(env.ENV_APP_SECRET),
tenant_id=os.environ.get(env.ENV_APP_TENANT),
)
cred_adapter = CredentialWrapper(credential)
client = ApplicationInsightsDataClient(cred_adapter)
client.config.headers.setdefault("x-api-key", os.environ.get(env.ENV_APPLICATION_INSIGHTS_KEY))
```
There should be a special Credential class for using ApplicationInsights without using wrappers and adding custom headers or ApplicationInsights should support authentication with ClientSecretCredentials only without requiring the Application Insights AppKey.
| 1.0 | ApplicationInsights does not work out of the box when using ClientSecretCredentials - - **Package Name**: azure-applicationinsights
- **Package Version**: 0.1.0
- **Operating System**: Linux
- **Python Version**: 3.8.7
**Describe the bug**
The applicationInsights client requires:
- the identity credential must have a signed_session (that is missing from the azure.identity.ClientSecretCredential) so a [wrapper](https://gist.github.com/lmazuel/cc683d82ea1d7b40208de7c9fc8de59d) has to be used to make it work.
- AppId and AppKey are needed to fetch queries succesfully: To pass the app key, the header needs to be modified in the applicationInsightDataClient
``` python
credential = ClientSecretCredential(
client_id=os.environ.get(env.ENV_APP_CLIENT_ID),
client_secret=os.environ.get(env.ENV_APP_SECRET),
tenant_id=os.environ.get(env.ENV_APP_TENANT),
)
cred_adapter = CredentialWrapper(credential)
client = ApplicationInsightsDataClient(cred_adapter)
client.config.headers.setdefault("x-api-key", os.environ.get(env.ENV_APPLICATION_INSIGHTS_KEY))
```
There should be a special Credential class for using ApplicationInsights without using wrappers and adding custom headers or ApplicationInsights should support authentication with ClientSecretCredentials only without requiring the Application Insights AppKey.
| non_defect | applicationinsights does not work out of the box when using clientsecretcredentials package name azure applicationinsights package version operating system linux python version describe the bug the applicationinsights client requires the identity credential must have a signed session that is missing from the azure identity clientsecretcredential so a has to be used to make it work appid and appkey are needed to fetch queries succesfully to pass the app key the header needs to be modified in the applicationinsightdataclient python credential clientsecretcredential client id os environ get env env app client id client secret os environ get env env app secret tenant id os environ get env env app tenant cred adapter credentialwrapper credential client applicationinsightsdataclient cred adapter client config headers setdefault x api key os environ get env env application insights key there should be a special credential class for using applicationinsights without using wrappers and adding custom headers or applicationinsights should support authentication with clientsecretcredentials only without requiring the application insights appkey | 0 |
16,685 | 2,931,469,148 | IssuesEvent | 2015-06-29 12:44:57 | MarcusWolschon/osmeditor4android | https://api.github.com/repos/MarcusWolschon/osmeditor4android | closed | 0.9.6 startup crash related to lock icon not being found | Defect FIXED! | A small number of crash dumps have been submitted that indicate an issue immediately after upgrade.
Issue is due to deprecated modes being turned on, which in itself is deprecated. | 1.0 | 0.9.6 startup crash related to lock icon not being found - A small number of crash dumps have been submitted that indicate an issue immediately after upgrade.
Issue is due to deprecated modes being turned on, which in itself is deprecated. | defect | startup crash related to lock icon not being found a small number of crash dumps have been submitted that indicate an issue immediately after upgrade issue is due to deprecated modes being turned on which in itself is deprecated | 1 |
159,936 | 6,064,642,121 | IssuesEvent | 2017-06-14 14:42:09 | openvstorage/openvstorage-health-check | https://api.github.com/repos/openvstorage/openvstorage-health-check | closed | healthcheck never ends when checks are waiting on memcache | priority_minor type_bug | Best solution is to use a timeout frame when the check is waiting on memcache.
| 1.0 | healthcheck never ends when checks are waiting on memcache - Best solution is to use a timeout frame when the check is waiting on memcache.
| non_defect | healthcheck never ends when checks are waiting on memcache best solution is to use a timeout frame when the check is waiting on memcache | 0 |
36,656 | 8,049,397,844 | IssuesEvent | 2018-08-01 09:58:05 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | TCP discovery with public-address specified doesn't handle correctly member shutdowns | Estimation: M Team: Core Type: Defect | I see "Connection reset by peer exceptions" in the (active) member log after one of Hazelcast members shuts down in my Docker environment.
### Environment
Docker host runs 2 containers with Hazelcast members. Both have configured TCP discovery and hosts public address. The Hazelcast port is exposed to the Docker host.
#### hazelcast.xml
```xml
<port auto-increment="false" port-count="100">5701</port>
<public-address>${hazelcast.public.ip}:${hazelcast.port}</public-address>
<join>
<multicast enabled="false">
</multicast>
<tcp-ip enabled="true">
<member>${hazelcast.public.ip}:5555</member>
<member>${hazelcast.public.ip}:5556</member>
</tcp-ip>
...
</join>
```
#### Start Docker containers
```bash
docker run -it --rm -p 5555:5701 hazelcast-tcp
docker run -it --rm -p 5556:5701 hazelcast-tcp
```
#### Start Hazelcast in the containers
Provide system properties used in `hazelcast.xml` (public Docker host IP address and the port - 5555 and 5556):
```
System.setProperty("hazelcast.public.ip", "192.168.1.105"); // replace the IP value with your own IP
System.setProperty("hazelcast.port", "5555"); // for the second container use 5556
hz = Hazelcast.newHazelcastInstance();
```
#### Shut down one of the containers
```java
hz.shutdown();
```
#### Check the log in the running container
The remaining cluster member try to connect to the stopped one and log contains bunch of entries similar to:
```
Oct 31, 2017 12:14:55 PM com.hazelcast.nio.tcp.TcpIpConnector
INFO: [192.168.1.105]:5555 [dev] [3.9] Connecting to /192.168.1.105:5556, timeout: 0, bind-any: true
Oct 31, 2017 12:14:55 PM com.hazelcast.nio.tcp.TcpIpConnectionManager
INFO: [192.168.1.105]:5555 [dev] [3.9] Established socket connection between /172.17.0.3:56203 and /192.168.1.105:5556
Oct 31, 2017 12:14:55 PM com.hazelcast.nio.tcp.TcpIpConnection
INFO: [192.168.1.105]:5555 [dev] [3.9] Connection[id=24, /172.17.0.3:56203->/192.168.1.105:5556, endpoint=[192.168.1.105]:5556, alive=false, type=NONE] closed. Reason: Connection closed by the other side
Oct 31, 2017 12:14:55 PM com.hazelcast.nio.tcp.TcpIpConnector
INFO: [192.168.1.105]:5555 [dev] [3.9] Connecting to /192.168.1.105:5556, timeout: 0, bind-any: true
Oct 31, 2017 12:14:55 PM com.hazelcast.nio.tcp.TcpIpConnectionManager
INFO: [192.168.1.105]:5555 [dev] [3.9] Established socket connection between /172.17.0.3:53561 and /192.168.1.105:5556
Oct 31, 2017 12:14:55 PM com.hazelcast.nio.tcp.TcpIpConnection
WARNING: [192.168.1.105]:5555 [dev] [3.9] Connection[id=25, /172.17.0.3:53561->/192.168.1.105:5556, endpoint=[192.168.1.105]:5556, alive=false, type=NONE] closed. Reason: Exception in Connection[id=25, /172.17.0.3:53561->/192.168.1.105:5556, endpoint=[192.168.1.105]:5556, alive=true, type=NONE], thread=hz._hzInstance_1_dev.IO.thread-in-0
java.io.IOException: Connection reset by peer
at sun.nio.ch.FileDispatcherImpl.read0(Native Method)
at sun.nio.ch.SocketDispatcher.read(SocketDispatcher.java:39)
at sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:223)
at sun.nio.ch.IOUtil.read(IOUtil.java:197)
at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:380)
at com.hazelcast.internal.networking.AbstractChannel.read(AbstractChannel.java:94)
at com.hazelcast.nio.tcp.MemberChannelInitializer.inboundProtocol(MemberChannelInitializer.java:101)
at com.hazelcast.nio.tcp.MemberChannelInitializer.initInbound(MemberChannelInitializer.java:73)
at com.hazelcast.internal.networking.nio.NioChannelReader.init(NioChannelReader.java:143)
at com.hazelcast.internal.networking.nio.NioChannelReader.handle(NioChannelReader.java:123)
at com.hazelcast.internal.networking.nio.NioThread.handleSelectionKey(NioThread.java:401)
at com.hazelcast.internal.networking.nio.NioThread.handleSelectionKeys(NioThread.java:386)
at com.hazelcast.internal.networking.nio.NioThread.selectLoop(NioThread.java:293)
at com.hazelcast.internal.networking.nio.NioThread.run(NioThread.java:248)
```
| 1.0 | TCP discovery with public-address specified doesn't handle correctly member shutdowns - I see "Connection reset by peer exceptions" in the (active) member log after one of Hazelcast members shuts down in my Docker environment.
### Environment
Docker host runs 2 containers with Hazelcast members. Both have configured TCP discovery and hosts public address. The Hazelcast port is exposed to the Docker host.
#### hazelcast.xml
```xml
<port auto-increment="false" port-count="100">5701</port>
<public-address>${hazelcast.public.ip}:${hazelcast.port}</public-address>
<join>
<multicast enabled="false">
</multicast>
<tcp-ip enabled="true">
<member>${hazelcast.public.ip}:5555</member>
<member>${hazelcast.public.ip}:5556</member>
</tcp-ip>
...
</join>
```
#### Start Docker containers
```bash
docker run -it --rm -p 5555:5701 hazelcast-tcp
docker run -it --rm -p 5556:5701 hazelcast-tcp
```
#### Start Hazelcast in the containers
Provide system properties used in `hazelcast.xml` (public Docker host IP address and the port - 5555 and 5556):
```
System.setProperty("hazelcast.public.ip", "192.168.1.105"); // replace the IP value with your own IP
System.setProperty("hazelcast.port", "5555"); // for the second container use 5556
hz = Hazelcast.newHazelcastInstance();
```
#### Shut down one of the containers
```java
hz.shutdown();
```
#### Check the log in the running container
The remaining cluster member try to connect to the stopped one and log contains bunch of entries similar to:
```
Oct 31, 2017 12:14:55 PM com.hazelcast.nio.tcp.TcpIpConnector
INFO: [192.168.1.105]:5555 [dev] [3.9] Connecting to /192.168.1.105:5556, timeout: 0, bind-any: true
Oct 31, 2017 12:14:55 PM com.hazelcast.nio.tcp.TcpIpConnectionManager
INFO: [192.168.1.105]:5555 [dev] [3.9] Established socket connection between /172.17.0.3:56203 and /192.168.1.105:5556
Oct 31, 2017 12:14:55 PM com.hazelcast.nio.tcp.TcpIpConnection
INFO: [192.168.1.105]:5555 [dev] [3.9] Connection[id=24, /172.17.0.3:56203->/192.168.1.105:5556, endpoint=[192.168.1.105]:5556, alive=false, type=NONE] closed. Reason: Connection closed by the other side
Oct 31, 2017 12:14:55 PM com.hazelcast.nio.tcp.TcpIpConnector
INFO: [192.168.1.105]:5555 [dev] [3.9] Connecting to /192.168.1.105:5556, timeout: 0, bind-any: true
Oct 31, 2017 12:14:55 PM com.hazelcast.nio.tcp.TcpIpConnectionManager
INFO: [192.168.1.105]:5555 [dev] [3.9] Established socket connection between /172.17.0.3:53561 and /192.168.1.105:5556
Oct 31, 2017 12:14:55 PM com.hazelcast.nio.tcp.TcpIpConnection
WARNING: [192.168.1.105]:5555 [dev] [3.9] Connection[id=25, /172.17.0.3:53561->/192.168.1.105:5556, endpoint=[192.168.1.105]:5556, alive=false, type=NONE] closed. Reason: Exception in Connection[id=25, /172.17.0.3:53561->/192.168.1.105:5556, endpoint=[192.168.1.105]:5556, alive=true, type=NONE], thread=hz._hzInstance_1_dev.IO.thread-in-0
java.io.IOException: Connection reset by peer
at sun.nio.ch.FileDispatcherImpl.read0(Native Method)
at sun.nio.ch.SocketDispatcher.read(SocketDispatcher.java:39)
at sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:223)
at sun.nio.ch.IOUtil.read(IOUtil.java:197)
at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:380)
at com.hazelcast.internal.networking.AbstractChannel.read(AbstractChannel.java:94)
at com.hazelcast.nio.tcp.MemberChannelInitializer.inboundProtocol(MemberChannelInitializer.java:101)
at com.hazelcast.nio.tcp.MemberChannelInitializer.initInbound(MemberChannelInitializer.java:73)
at com.hazelcast.internal.networking.nio.NioChannelReader.init(NioChannelReader.java:143)
at com.hazelcast.internal.networking.nio.NioChannelReader.handle(NioChannelReader.java:123)
at com.hazelcast.internal.networking.nio.NioThread.handleSelectionKey(NioThread.java:401)
at com.hazelcast.internal.networking.nio.NioThread.handleSelectionKeys(NioThread.java:386)
at com.hazelcast.internal.networking.nio.NioThread.selectLoop(NioThread.java:293)
at com.hazelcast.internal.networking.nio.NioThread.run(NioThread.java:248)
```
| defect | tcp discovery with public address specified doesn t handle correctly member shutdowns i see connection reset by peer exceptions in the active member log after one of hazelcast members shuts down in my docker environment environment docker host runs containers with hazelcast members both have configured tcp discovery and hosts public address the hazelcast port is exposed to the docker host hazelcast xml xml hazelcast public ip hazelcast port hazelcast public ip hazelcast public ip start docker containers bash docker run it rm p hazelcast tcp docker run it rm p hazelcast tcp start hazelcast in the containers provide system properties used in hazelcast xml public docker host ip address and the port and system setproperty hazelcast public ip replace the ip value with your own ip system setproperty hazelcast port for the second container use hz hazelcast newhazelcastinstance shut down one of the containers java hz shutdown check the log in the running container the remaining cluster member try to connect to the stopped one and log contains bunch of entries similar to oct pm com hazelcast nio tcp tcpipconnector info connecting to timeout bind any true oct pm com hazelcast nio tcp tcpipconnectionmanager info established socket connection between and oct pm com hazelcast nio tcp tcpipconnection info connection alive false type none closed reason connection closed by the other side oct pm com hazelcast nio tcp tcpipconnector info connecting to timeout bind any true oct pm com hazelcast nio tcp tcpipconnectionmanager info established socket connection between and oct pm com hazelcast nio tcp tcpipconnection warning connection alive false type none closed reason exception in connection alive true type none thread hz hzinstance dev io thread in java io ioexception connection reset by peer at sun nio ch filedispatcherimpl native method at sun nio ch socketdispatcher read socketdispatcher java at sun nio ch ioutil readintonativebuffer ioutil java at sun nio ch ioutil read ioutil java at sun nio ch socketchannelimpl read socketchannelimpl java at com hazelcast internal networking abstractchannel read abstractchannel java at com hazelcast nio tcp memberchannelinitializer inboundprotocol memberchannelinitializer java at com hazelcast nio tcp memberchannelinitializer initinbound memberchannelinitializer java at com hazelcast internal networking nio niochannelreader init niochannelreader java at com hazelcast internal networking nio niochannelreader handle niochannelreader java at com hazelcast internal networking nio niothread handleselectionkey niothread java at com hazelcast internal networking nio niothread handleselectionkeys niothread java at com hazelcast internal networking nio niothread selectloop niothread java at com hazelcast internal networking nio niothread run niothread java | 1 |
484,282 | 13,937,402,661 | IssuesEvent | 2020-10-22 14:06:34 | ViRGiL175/java-diner-automation | https://api.github.com/repos/ViRGiL175/java-diner-automation | opened | Босс видит фидбек клиентов через интерфейс | complexity:_3 priority:_1 type:_story | Босс хочет наблюдать за фидбеком клиентов дайнера через удобный и понятный интерфейс приложения, чтобы знать, насколько клиенты довольны обслуживанием
## Описание
> Описание задачи, схемы, картинки и т.д.
## Критерии выполненности
- [ ] Босс видит фидбек клиентов
- [ ] Данные регулярно обновляются
## Связанное
> Всяческие подробности, ссылки, документация и т.д.
| 1.0 | Босс видит фидбек клиентов через интерфейс - Босс хочет наблюдать за фидбеком клиентов дайнера через удобный и понятный интерфейс приложения, чтобы знать, насколько клиенты довольны обслуживанием
## Описание
> Описание задачи, схемы, картинки и т.д.
## Критерии выполненности
- [ ] Босс видит фидбек клиентов
- [ ] Данные регулярно обновляются
## Связанное
> Всяческие подробности, ссылки, документация и т.д.
| non_defect | босс видит фидбек клиентов через интерфейс босс хочет наблюдать за фидбеком клиентов дайнера через удобный и понятный интерфейс приложения чтобы знать насколько клиенты довольны обслуживанием описание описание задачи схемы картинки и т д критерии выполненности босс видит фидбек клиентов данные регулярно обновляются связанное всяческие подробности ссылки документация и т д | 0 |
66,248 | 20,103,291,856 | IssuesEvent | 2022-02-07 07:54:56 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | Compact GenericRecords' getFieldKind should throw [HZ-879] | Type: Defect Team: Client Source: Internal Estimation: S Module: Serialization to-jira | When field name does not exist in generic record according to this code: https://github.com/hazelcast/hazelcast/blob/0ab9eb067f04cbbb62d8998ee21c99cb05239761/hazelcast/src/main/java/com/hazelcast/nio/serialization/GenericRecord.java#L119-L125
getFieldKind should throw and should not return null. (NonNull annotation)
However, as can be seen from https://github.com/hazelcast/hazelcast/blob/77ef2bfdfb19005b560cefb2a104994019983d2a/hazelcast/src/main/java/com/hazelcast/internal/serialization/impl/compact/DeserializedGenericRecord.java#L107-L111 and https://github.com/hazelcast/hazelcast/blob/4544e4a6fa5060fa9ae2984a20ed3956099cc583/hazelcast/src/main/java/com/hazelcast/internal/serialization/impl/compact/CompactInternalGenericRecord.java#L165-L169 they don't throw.
| 1.0 | Compact GenericRecords' getFieldKind should throw [HZ-879] - When field name does not exist in generic record according to this code: https://github.com/hazelcast/hazelcast/blob/0ab9eb067f04cbbb62d8998ee21c99cb05239761/hazelcast/src/main/java/com/hazelcast/nio/serialization/GenericRecord.java#L119-L125
getFieldKind should throw and should not return null. (NonNull annotation)
However, as can be seen from https://github.com/hazelcast/hazelcast/blob/77ef2bfdfb19005b560cefb2a104994019983d2a/hazelcast/src/main/java/com/hazelcast/internal/serialization/impl/compact/DeserializedGenericRecord.java#L107-L111 and https://github.com/hazelcast/hazelcast/blob/4544e4a6fa5060fa9ae2984a20ed3956099cc583/hazelcast/src/main/java/com/hazelcast/internal/serialization/impl/compact/CompactInternalGenericRecord.java#L165-L169 they don't throw.
| defect | compact genericrecords getfieldkind should throw when field name does not exist in generic record according to this code getfieldkind should throw and should not return null nonnull annotation however as can be seen from and they don t throw | 1 |
36,687 | 12,418,576,432 | IssuesEvent | 2020-05-23 01:02:00 | wrbejar/Autofac | https://api.github.com/repos/wrbejar/Autofac | opened | CVE-2020-7656 (Medium) detected in jquery-1.4.1.min.js, jquery-1.4.1.js | security vulnerability | ## CVE-2020-7656 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-1.4.1.min.js</b>, <b>jquery-1.4.1.js</b></p></summary>
<p>
<details><summary><b>jquery-1.4.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.1/jquery.min.js</a></p>
<p>Path to vulnerable library: /Autofac/src/MultitenantExample.MvcApplication/Scripts/jquery-1.4.1.min.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.4.1.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.4.1.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.1/jquery.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.1/jquery.js</a></p>
<p>Path to vulnerable library: /Autofac/src/MultitenantExample.MvcApplication/Scripts/jquery-1.4.1.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.4.1.js** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jquery prior to 1.9.0 allows Cross-site Scripting attacks via the load method. The load method fails to recognize and remove "<script>" HTML tags that contain a whitespace character, i.e: "</script >", which results in the enclosed script logic to be executed.
<p>Publish Date: 2020-05-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7656>CVE-2020-7656</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7656">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7656</a></p>
<p>Release Date: 2020-05-19</p>
<p>Fix Resolution: 1.9.0b1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.4.1","isTransitiveDependency":false,"dependencyTree":"jquery:1.4.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.9.0b1"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.4.1","isTransitiveDependency":false,"dependencyTree":"jquery:1.4.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.9.0b1"}],"vulnerabilityIdentifier":"CVE-2020-7656","vulnerabilityDetails":"jquery prior to 1.9.0 allows Cross-site Scripting attacks via the load method. The load method fails to recognize and remove \"\u003cscript\u003e\" HTML tags that contain a whitespace character, i.e: \"\u003c/script \u003e\", which results in the enclosed script logic to be executed.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7656","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | True | CVE-2020-7656 (Medium) detected in jquery-1.4.1.min.js, jquery-1.4.1.js - ## CVE-2020-7656 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-1.4.1.min.js</b>, <b>jquery-1.4.1.js</b></p></summary>
<p>
<details><summary><b>jquery-1.4.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.1/jquery.min.js</a></p>
<p>Path to vulnerable library: /Autofac/src/MultitenantExample.MvcApplication/Scripts/jquery-1.4.1.min.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.4.1.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.4.1.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.1/jquery.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.4.1/jquery.js</a></p>
<p>Path to vulnerable library: /Autofac/src/MultitenantExample.MvcApplication/Scripts/jquery-1.4.1.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.4.1.js** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jquery prior to 1.9.0 allows Cross-site Scripting attacks via the load method. The load method fails to recognize and remove "<script>" HTML tags that contain a whitespace character, i.e: "</script >", which results in the enclosed script logic to be executed.
<p>Publish Date: 2020-05-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7656>CVE-2020-7656</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7656">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-7656</a></p>
<p>Release Date: 2020-05-19</p>
<p>Fix Resolution: 1.9.0b1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.4.1","isTransitiveDependency":false,"dependencyTree":"jquery:1.4.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.9.0b1"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.4.1","isTransitiveDependency":false,"dependencyTree":"jquery:1.4.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.9.0b1"}],"vulnerabilityIdentifier":"CVE-2020-7656","vulnerabilityDetails":"jquery prior to 1.9.0 allows Cross-site Scripting attacks via the load method. The load method fails to recognize and remove \"\u003cscript\u003e\" HTML tags that contain a whitespace character, i.e: \"\u003c/script \u003e\", which results in the enclosed script logic to be executed.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7656","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | non_defect | cve medium detected in jquery min js jquery js cve medium severity vulnerability vulnerable libraries jquery min js jquery js jquery min js javascript library for dom operations library home page a href path to vulnerable library autofac src multitenantexample mvcapplication scripts jquery min js dependency hierarchy x jquery min js vulnerable library jquery js javascript library for dom operations library home page a href path to vulnerable library autofac src multitenantexample mvcapplication scripts jquery js dependency hierarchy x jquery js vulnerable library vulnerability details jquery prior to allows cross site scripting attacks via the load method the load method fails to recognize and remove html tags that contain a whitespace character i e which results in the enclosed script logic to be executed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails jquery prior to allows cross site scripting attacks via the load method the load method fails to recognize and remove html tags that contain a whitespace character i e script which results in the enclosed script logic to be executed vulnerabilityurl | 0 |
44,399 | 12,132,757,919 | IssuesEvent | 2020-04-23 07:52:33 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | opened | hazelcast-client:4.0 is missing | Type: Defect | Looks like the hazelcast-client 4.X was not released. For the Hazelcast server and Hazelcast-hibernate5 artefacts we can migrate to Hazelcast 4 but there is not client artefact at Maven central (or JCenter). [Your documentation](https://hazelcast.org/imdg/clients-languages/java/) already mentions version 4.0.1 of the hazelcast-client but I can not find it anywhere. | 1.0 | hazelcast-client:4.0 is missing - Looks like the hazelcast-client 4.X was not released. For the Hazelcast server and Hazelcast-hibernate5 artefacts we can migrate to Hazelcast 4 but there is not client artefact at Maven central (or JCenter). [Your documentation](https://hazelcast.org/imdg/clients-languages/java/) already mentions version 4.0.1 of the hazelcast-client but I can not find it anywhere. | defect | hazelcast client is missing looks like the hazelcast client x was not released for the hazelcast server and hazelcast artefacts we can migrate to hazelcast but there is not client artefact at maven central or jcenter already mentions version of the hazelcast client but i can not find it anywhere | 1 |
44,915 | 12,440,273,664 | IssuesEvent | 2020-05-26 11:41:21 | primefaces/primefaces | https://api.github.com/repos/primefaces/primefaces | closed | Native tooltip [title] does not change for FileUpload component in advanced mode | defect | By default native HTML `input type="file"` will have a tolltip specific to browser.
On Firefox its: _No file selected._
On chrome its: _No file chosen_
After a file has been selected by user, the tolltip will change to selected file name.
On Primefaces FileUpload with Basic mode this works just as expected.
But when used as Advanced mode the tooltip will never change, it will always be same.
Apart from this issue, it would be good to have tooltip attribute for Fileupload component.
| 1.0 | Native tooltip [title] does not change for FileUpload component in advanced mode - By default native HTML `input type="file"` will have a tolltip specific to browser.
On Firefox its: _No file selected._
On chrome its: _No file chosen_
After a file has been selected by user, the tolltip will change to selected file name.
On Primefaces FileUpload with Basic mode this works just as expected.
But when used as Advanced mode the tooltip will never change, it will always be same.
Apart from this issue, it would be good to have tooltip attribute for Fileupload component.
| defect | native tooltip does not change for fileupload component in advanced mode by default native html input type file will have a tolltip specific to browser on firefox its no file selected on chrome its no file chosen after a file has been selected by user the tolltip will change to selected file name on primefaces fileupload with basic mode this works just as expected but when used as advanced mode the tooltip will never change it will always be same apart from this issue it would be good to have tooltip attribute for fileupload component | 1 |
139,262 | 20,820,627,113 | IssuesEvent | 2022-03-18 15:01:47 | cuzekn/My-Best-Book | https://api.github.com/repos/cuzekn/My-Best-Book | opened | ユーザー認証(ユーザー作成画面とログイン画面)を実装 | design | [やること]
- Reactアプリの環境構築
- ルーティングの作成
- ユーザー作成画面の作成(/signup)
URL例) localhost/signup
- ログイン画面の作成(/login)
URL例) localhost/login | 1.0 | ユーザー認証(ユーザー作成画面とログイン画面)を実装 - [やること]
- Reactアプリの環境構築
- ルーティングの作成
- ユーザー作成画面の作成(/signup)
URL例) localhost/signup
- ログイン画面の作成(/login)
URL例) localhost/login | non_defect | ユーザー認証 ユーザー作成画面とログイン画面 を実装 reactアプリの環境構築 ルーティングの作成 ユーザー作成画面の作成 signup url例 localhost signup ログイン画面の作成 login url例 localhost login | 0 |
30,606 | 6,192,417,563 | IssuesEvent | 2017-07-05 01:38:53 | cakephp/cakephp | https://api.github.com/repos/cakephp/cakephp | closed | translate behaviour doesn't seem to patch translations on edit | behaviors Defect i18n On hold | This is a (multiple allowed):
* [x] bug
* [ ] enhancement
* [ ] feature-discussion (RFC)
* CakePHP Version: 3.4.7
* Platform and Target: ubuntu 16.04
### What you did
I created form fields using as follows:
`_translations.en_GB.name`
etc..
according to the docs here:
https://book.cakephp.org/3.0/en/orm/behaviors/translate.html#saving-multiple-translations
### What happened
when I use
```php
$entity = $this->ModelTable->find('translations')->where(['id'=>'1]->first();
$this->ModelTable->patchEntity($entity, $this->request->getData(), ['translations'=>true]);
```
translated fields are not patched even though the data from the form seems to be correct.
### What you expected to happen
I would expect translated fields to be patched as well
| 1.0 | translate behaviour doesn't seem to patch translations on edit - This is a (multiple allowed):
* [x] bug
* [ ] enhancement
* [ ] feature-discussion (RFC)
* CakePHP Version: 3.4.7
* Platform and Target: ubuntu 16.04
### What you did
I created form fields using as follows:
`_translations.en_GB.name`
etc..
according to the docs here:
https://book.cakephp.org/3.0/en/orm/behaviors/translate.html#saving-multiple-translations
### What happened
when I use
```php
$entity = $this->ModelTable->find('translations')->where(['id'=>'1]->first();
$this->ModelTable->patchEntity($entity, $this->request->getData(), ['translations'=>true]);
```
translated fields are not patched even though the data from the form seems to be correct.
### What you expected to happen
I would expect translated fields to be patched as well
| defect | translate behaviour doesn t seem to patch translations on edit this is a multiple allowed bug enhancement feature discussion rfc cakephp version platform and target ubuntu what you did i created form fields using as follows translations en gb name etc according to the docs here what happened when i use php entity this modeltable find translations where first this modeltable patchentity entity this request getdata translated fields are not patched even though the data from the form seems to be correct what you expected to happen i would expect translated fields to be patched as well | 1 |
71,147 | 23,469,347,357 | IssuesEvent | 2022-08-16 20:04:02 | department-of-veterans-affairs/va.gov-cms | https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms | opened | Duplicate calls to action on Vet Centers Dashboard cards | Needs refining ⭐️ Sitewide CMS 508/Accessibility 508-defect-4 | ## Description
On the Vet Center Dashboard in the CMS, there are multiple cards that each have a call to action for "View" and "Edit", causing duplicate links. Ideally each link should have distinct, unique text to indicate to the user where the link would direct them - "Edit Locations Page" or "View Main Vet Center Page"
## Screenshot

## Accessibility Standard
WCAG version 2.0 AAA, [Criterion 2.4.9](https://www.w3.org/WAI/WCAG21/Understanding/link-purpose-link-only.html)
## Acceptance Criteria
- [ ] UX review to determine content updates for links
- [ ] Technical review
- [ ] Change management consulted
- [ ] Implementation ticket created
### CMS Team
Please check the team(s) that will do this work.
- [ ] `Program`
- [ ] `Platform CMS Team`
- [ ] `Sitewide Crew`
- [ ] `⭐️ Sitewide CMS`
- [ ] `⭐️ Public Websites`
- [ ] `⭐️ Facilities`
- [ ] `⭐️ User support`
| 1.0 | Duplicate calls to action on Vet Centers Dashboard cards - ## Description
On the Vet Center Dashboard in the CMS, there are multiple cards that each have a call to action for "View" and "Edit", causing duplicate links. Ideally each link should have distinct, unique text to indicate to the user where the link would direct them - "Edit Locations Page" or "View Main Vet Center Page"
## Screenshot

## Accessibility Standard
WCAG version 2.0 AAA, [Criterion 2.4.9](https://www.w3.org/WAI/WCAG21/Understanding/link-purpose-link-only.html)
## Acceptance Criteria
- [ ] UX review to determine content updates for links
- [ ] Technical review
- [ ] Change management consulted
- [ ] Implementation ticket created
### CMS Team
Please check the team(s) that will do this work.
- [ ] `Program`
- [ ] `Platform CMS Team`
- [ ] `Sitewide Crew`
- [ ] `⭐️ Sitewide CMS`
- [ ] `⭐️ Public Websites`
- [ ] `⭐️ Facilities`
- [ ] `⭐️ User support`
| defect | duplicate calls to action on vet centers dashboard cards description on the vet center dashboard in the cms there are multiple cards that each have a call to action for view and edit causing duplicate links ideally each link should have distinct unique text to indicate to the user where the link would direct them edit locations page or view main vet center page screenshot accessibility standard wcag version aaa acceptance criteria ux review to determine content updates for links technical review change management consulted implementation ticket created cms team please check the team s that will do this work program platform cms team sitewide crew ⭐️ sitewide cms ⭐️ public websites ⭐️ facilities ⭐️ user support | 1 |
331,009 | 10,058,746,177 | IssuesEvent | 2019-07-22 14:31:18 | trailofbits/manticore | https://api.github.com/repos/trailofbits/manticore | opened | Assertion failure: assert value.size == self.size when gas is ignored | ethereum help wanted priority-high | ### OS / Environment
Ubuntu 18.04
### Manticore version
cc7811b10d9ce4f5c3444ebceff16b3ed2753f21
### Python version
Python 3.6
### Summary of the problem
Manticode crashes with an assertion failure when gas computation is ignored.
### Step to reproduce the behavior
Use this solidity contract:
```
contract C {
function hash(string x) public returns (bytes32) {
return(keccak256(x));
}
}
```
with manticore:
```
$ manticore bug.sol --txlimit 1 --evm.oog ignore
```
### Expected behavior
It should should not crash. It's unclear how to proceed with the symbolic exploration since manticore will waste resourced exploring the `keccak256` code. Maybe a warning saying that such path will not be explored?
### Actual behavior
```
2019-07-22 10:24:54,158: [15617] m.main:INFO: Registered plugins: DetectReentrancyAdvanced, DetectSuicidal, DetectUnusedRetVal, DetectInvalid, DetectUninitializedMemory, DetectIntegerOverflow, DetectEnvInstruction, DetectExternalCallAndLeak, DetectDelegatecall, DetectReentrancySimple, DetectUninitializedStorage
2019-07-22 10:24:54,158: [15617] m.main:INFO: Beginning analysis
2019-07-22 10:24:54,163: [15617] m.e.manticore:INFO: Starting symbolic create contract
2019-07-22 10:24:54,710: [15617] m.e.manticore:INFO: Starting symbolic transaction: 0
2019-07-22 10:25:10,260: [15916] m.c.worker:ERROR: Exception in state 0: AssertionError()
Traceback (most recent call last):
File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/core/worker.py", line 121, in run
current_state.execute()
File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/ethereum/state.py", line 6, in execute
return self._platform.execute()
File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/platforms/evm.py", line 2684, in execute
self.current_vm.execute()
File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/platforms/evm.py", line 1210, in execute
last_pc, last_gas, instruction, arguments, fee, allocated = self._checkpoint()
File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/platforms/evm.py", line 1106, in _checkpoint
fee = self._calculate_gas(*arguments)
File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/platforms/evm.py", line 1079, in _calculate_gas
return current.fee + implementation(*arguments)
File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/platforms/evm.py", line 1533, in SHA3_gas
return GSHA3WORD * (ceil32(size) // 32) + memfee
File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/core/smtlib/expression.py", line 295, in __add__
return BitVecAdd(self, self.cast(other))
File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/core/smtlib/expression.py", line 284, in cast
assert value.size == self.size
AssertionError
2019-07-22 10:25:10,346: [15617] m.e.manticore:INFO: 0 alive states, 3 terminated states
2019-07-22 10:25:10,378: [16754] m.c.manticore:INFO: Generated testcase No. 0 - RETURN(1 txs)
2019-07-22 10:25:10,383: [16755] m.c.manticore:INFO: Generated testcase No. 1 - REVERT(2 txs)
2019-07-22 10:25:10,387: [16757] m.c.manticore:INFO: Generated testcase No. 2 - REVERT(2 txs)
2019-07-22 10:25:10,417: [16760] m.c.manticore:INFO: Generated testcase No. 3 - RETURN(1 txs)
```
| 1.0 | Assertion failure: assert value.size == self.size when gas is ignored - ### OS / Environment
Ubuntu 18.04
### Manticore version
cc7811b10d9ce4f5c3444ebceff16b3ed2753f21
### Python version
Python 3.6
### Summary of the problem
Manticode crashes with an assertion failure when gas computation is ignored.
### Step to reproduce the behavior
Use this solidity contract:
```
contract C {
function hash(string x) public returns (bytes32) {
return(keccak256(x));
}
}
```
with manticore:
```
$ manticore bug.sol --txlimit 1 --evm.oog ignore
```
### Expected behavior
It should should not crash. It's unclear how to proceed with the symbolic exploration since manticore will waste resourced exploring the `keccak256` code. Maybe a warning saying that such path will not be explored?
### Actual behavior
```
2019-07-22 10:24:54,158: [15617] m.main:INFO: Registered plugins: DetectReentrancyAdvanced, DetectSuicidal, DetectUnusedRetVal, DetectInvalid, DetectUninitializedMemory, DetectIntegerOverflow, DetectEnvInstruction, DetectExternalCallAndLeak, DetectDelegatecall, DetectReentrancySimple, DetectUninitializedStorage
2019-07-22 10:24:54,158: [15617] m.main:INFO: Beginning analysis
2019-07-22 10:24:54,163: [15617] m.e.manticore:INFO: Starting symbolic create contract
2019-07-22 10:24:54,710: [15617] m.e.manticore:INFO: Starting symbolic transaction: 0
2019-07-22 10:25:10,260: [15916] m.c.worker:ERROR: Exception in state 0: AssertionError()
Traceback (most recent call last):
File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/core/worker.py", line 121, in run
current_state.execute()
File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/ethereum/state.py", line 6, in execute
return self._platform.execute()
File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/platforms/evm.py", line 2684, in execute
self.current_vm.execute()
File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/platforms/evm.py", line 1210, in execute
last_pc, last_gas, instruction, arguments, fee, allocated = self._checkpoint()
File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/platforms/evm.py", line 1106, in _checkpoint
fee = self._calculate_gas(*arguments)
File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/platforms/evm.py", line 1079, in _calculate_gas
return current.fee + implementation(*arguments)
File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/platforms/evm.py", line 1533, in SHA3_gas
return GSHA3WORD * (ceil32(size) // 32) + memfee
File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/core/smtlib/expression.py", line 295, in __add__
return BitVecAdd(self, self.cast(other))
File "/home/gustavo/.local/lib/python3.6/site-packages/manticore-0.3.0-py3.6.egg/manticore/core/smtlib/expression.py", line 284, in cast
assert value.size == self.size
AssertionError
2019-07-22 10:25:10,346: [15617] m.e.manticore:INFO: 0 alive states, 3 terminated states
2019-07-22 10:25:10,378: [16754] m.c.manticore:INFO: Generated testcase No. 0 - RETURN(1 txs)
2019-07-22 10:25:10,383: [16755] m.c.manticore:INFO: Generated testcase No. 1 - REVERT(2 txs)
2019-07-22 10:25:10,387: [16757] m.c.manticore:INFO: Generated testcase No. 2 - REVERT(2 txs)
2019-07-22 10:25:10,417: [16760] m.c.manticore:INFO: Generated testcase No. 3 - RETURN(1 txs)
```
| non_defect | assertion failure assert value size self size when gas is ignored os environment ubuntu manticore version python version python summary of the problem manticode crashes with an assertion failure when gas computation is ignored step to reproduce the behavior use this solidity contract contract c function hash string x public returns return x with manticore manticore bug sol txlimit evm oog ignore expected behavior it should should not crash it s unclear how to proceed with the symbolic exploration since manticore will waste resourced exploring the code maybe a warning saying that such path will not be explored actual behavior m main info registered plugins detectreentrancyadvanced detectsuicidal detectunusedretval detectinvalid detectuninitializedmemory detectintegeroverflow detectenvinstruction detectexternalcallandleak detectdelegatecall detectreentrancysimple detectuninitializedstorage m main info beginning analysis m e manticore info starting symbolic create contract m e manticore info starting symbolic transaction m c worker error exception in state assertionerror traceback most recent call last file home gustavo local lib site packages manticore egg manticore core worker py line in run current state execute file home gustavo local lib site packages manticore egg manticore ethereum state py line in execute return self platform execute file home gustavo local lib site packages manticore egg manticore platforms evm py line in execute self current vm execute file home gustavo local lib site packages manticore egg manticore platforms evm py line in execute last pc last gas instruction arguments fee allocated self checkpoint file home gustavo local lib site packages manticore egg manticore platforms evm py line in checkpoint fee self calculate gas arguments file home gustavo local lib site packages manticore egg manticore platforms evm py line in calculate gas return current fee implementation arguments file home gustavo local lib site packages manticore egg manticore platforms evm py line in gas return size memfee file home gustavo local lib site packages manticore egg manticore core smtlib expression py line in add return bitvecadd self self cast other file home gustavo local lib site packages manticore egg manticore core smtlib expression py line in cast assert value size self size assertionerror m e manticore info alive states terminated states m c manticore info generated testcase no return txs m c manticore info generated testcase no revert txs m c manticore info generated testcase no revert txs m c manticore info generated testcase no return txs | 0 |
204,769 | 23,279,881,428 | IssuesEvent | 2022-08-05 10:57:00 | Satheesh575555/Openssl_1_1_0_CVE-2021-23841 | https://api.github.com/repos/Satheesh575555/Openssl_1_1_0_CVE-2021-23841 | opened | CVE-2017-3738 (Medium) detected in opensslOpenSSL_1_1_0g | security vulnerability | ## CVE-2017-3738 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>opensslOpenSSL_1_1_0g</b></p></summary>
<p>
<p>Akamai fork of openssl master.</p>
<p>Library home page: <a href=https://github.com/akamai/openssl.git>https://github.com/akamai/openssl.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Satheesh575555/Openssl_1_1_0_CVE-2021-23841/commit/4c0cda99bc44e45f32ea15e98aeb0a895ad1dd56">4c0cda99bc44e45f32ea15e98aeb0a895ad1dd56</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/crypto/bn/asm/rsaz-avx2.pl</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
There is an overflow bug in the AVX2 Montgomery multiplication procedure used in exponentiation with 1024-bit moduli. No EC algorithms are affected. Analysis suggests that attacks against RSA and DSA as a result of this defect would be very difficult to perform and are not believed likely. Attacks against DH1024 are considered just feasible, because most of the work necessary to deduce information about a private key may be performed offline. The amount of resources required for such an attack would be significant. However, for an attack on TLS to be meaningful, the server would have to share the DH1024 private key among multiple clients, which is no longer an option since CVE-2016-0701. This only affects processors that support the AVX2 but not ADX extensions like Intel Haswell (4th generation). Note: The impact from this issue is similar to CVE-2017-3736, CVE-2017-3732 and CVE-2015-3193. OpenSSL version 1.0.2-1.0.2m and 1.1.0-1.1.0g are affected. Fixed in OpenSSL 1.0.2n. Due to the low severity of this issue we are not issuing a new release of OpenSSL 1.1.0 at this time. The fix will be included in OpenSSL 1.1.0h when it becomes available. The fix is also available in commit e502cc86d in the OpenSSL git repository.
<p>Publish Date: 2017-12-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-3738>CVE-2017-3738</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://gitlab.alpinelinux.org/alpine/aports/issues/8273">https://gitlab.alpinelinux.org/alpine/aports/issues/8273</a></p>
<p>Release Date: 2017-12-07</p>
<p>Fix Resolution: 1.0.2n</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2017-3738 (Medium) detected in opensslOpenSSL_1_1_0g - ## CVE-2017-3738 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>opensslOpenSSL_1_1_0g</b></p></summary>
<p>
<p>Akamai fork of openssl master.</p>
<p>Library home page: <a href=https://github.com/akamai/openssl.git>https://github.com/akamai/openssl.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Satheesh575555/Openssl_1_1_0_CVE-2021-23841/commit/4c0cda99bc44e45f32ea15e98aeb0a895ad1dd56">4c0cda99bc44e45f32ea15e98aeb0a895ad1dd56</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/crypto/bn/asm/rsaz-avx2.pl</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
There is an overflow bug in the AVX2 Montgomery multiplication procedure used in exponentiation with 1024-bit moduli. No EC algorithms are affected. Analysis suggests that attacks against RSA and DSA as a result of this defect would be very difficult to perform and are not believed likely. Attacks against DH1024 are considered just feasible, because most of the work necessary to deduce information about a private key may be performed offline. The amount of resources required for such an attack would be significant. However, for an attack on TLS to be meaningful, the server would have to share the DH1024 private key among multiple clients, which is no longer an option since CVE-2016-0701. This only affects processors that support the AVX2 but not ADX extensions like Intel Haswell (4th generation). Note: The impact from this issue is similar to CVE-2017-3736, CVE-2017-3732 and CVE-2015-3193. OpenSSL version 1.0.2-1.0.2m and 1.1.0-1.1.0g are affected. Fixed in OpenSSL 1.0.2n. Due to the low severity of this issue we are not issuing a new release of OpenSSL 1.1.0 at this time. The fix will be included in OpenSSL 1.1.0h when it becomes available. The fix is also available in commit e502cc86d in the OpenSSL git repository.
<p>Publish Date: 2017-12-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-3738>CVE-2017-3738</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://gitlab.alpinelinux.org/alpine/aports/issues/8273">https://gitlab.alpinelinux.org/alpine/aports/issues/8273</a></p>
<p>Release Date: 2017-12-07</p>
<p>Fix Resolution: 1.0.2n</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_defect | cve medium detected in opensslopenssl cve medium severity vulnerability vulnerable library opensslopenssl akamai fork of openssl master library home page a href found in head commit a href found in base branch master vulnerable source files crypto bn asm rsaz pl vulnerability details there is an overflow bug in the montgomery multiplication procedure used in exponentiation with bit moduli no ec algorithms are affected analysis suggests that attacks against rsa and dsa as a result of this defect would be very difficult to perform and are not believed likely attacks against are considered just feasible because most of the work necessary to deduce information about a private key may be performed offline the amount of resources required for such an attack would be significant however for an attack on tls to be meaningful the server would have to share the private key among multiple clients which is no longer an option since cve this only affects processors that support the but not adx extensions like intel haswell generation note the impact from this issue is similar to cve cve and cve openssl version and are affected fixed in openssl due to the low severity of this issue we are not issuing a new release of openssl at this time the fix will be included in openssl when it becomes available the fix is also available in commit in the openssl git repository publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
40,798 | 16,546,026,188 | IssuesEvent | 2021-05-28 00:11:34 | microsoft/vscode-cpptools | https://api.github.com/repos/microsoft/vscode-cpptools | closed | Wrong bitness for xtensa-esp32-elf-gcc | Feature: Configuration Language Service bug fixed (release pending) quick fix | I'm also having the same issue with GCC 8 and Xtensa (ESP32):

This is obviously wrong, given that size_t IS unsigned int, and IntelliSense known that:

Everything else works fine, except standard literals for strings (like `s` and `sv`), which are broken in the same way.
Needless to say, this code compiles fine with GCC 8.4, triple is xtensa-esp32-elf, `sizeof(int) == 4`. The C/C++ extension I'm using is version 1.2.1.
_Originally posted by @mcilloni in https://github.com/microsoft/vscode-cpptools/issues/6573#issuecomment-780442797_ | 1.0 | Wrong bitness for xtensa-esp32-elf-gcc - I'm also having the same issue with GCC 8 and Xtensa (ESP32):

This is obviously wrong, given that size_t IS unsigned int, and IntelliSense known that:

Everything else works fine, except standard literals for strings (like `s` and `sv`), which are broken in the same way.
Needless to say, this code compiles fine with GCC 8.4, triple is xtensa-esp32-elf, `sizeof(int) == 4`. The C/C++ extension I'm using is version 1.2.1.
_Originally posted by @mcilloni in https://github.com/microsoft/vscode-cpptools/issues/6573#issuecomment-780442797_ | non_defect | wrong bitness for xtensa elf gcc i m also having the same issue with gcc and xtensa this is obviously wrong given that size t is unsigned int and intellisense known that everything else works fine except standard literals for strings like s and sv which are broken in the same way needless to say this code compiles fine with gcc triple is xtensa elf sizeof int the c c extension i m using is version originally posted by mcilloni in | 0 |
74,035 | 15,298,939,448 | IssuesEvent | 2021-02-24 10:19:47 | rsoreq/kendo-ui-core | https://api.github.com/repos/rsoreq/kendo-ui-core | opened | CVE-2016-10531 (Medium) detected in marked-0.3.19.js | security vulnerability | ## CVE-2016-10531 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>marked-0.3.19.js</b></p></summary>
<p>A markdown parser built for speed</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/marked/0.3.19/marked.js">https://cdnjs.cloudflare.com/ajax/libs/marked/0.3.19/marked.js</a></p>
<p>Path to dependency file: kendo-ui-core/docs/node_modules/marked/www/demo.html</p>
<p>Path to vulnerable library: kendo-ui-core/docs/node_modules/marked/www/../lib/marked.js,kendo-ui-core/docs-aspnet/node_modules/marked/www/../lib/marked.js</p>
<p>
Dependency Hierarchy:
- :x: **marked-0.3.19.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/rsoreq/kendo-ui-core/commit/62afbcdf79c4c7052417ecc86eb31bd6bc04e1ad">62afbcdf79c4c7052417ecc86eb31bd6bc04e1ad</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
marked is an application that is meant to parse and compile markdown. Due to the way that marked 0.3.5 and earlier parses input, specifically HTML entities, it's possible to bypass marked's content injection protection (`sanitize: true`) to inject a `javascript:` URL. This flaw exists because `&#xNNanything;` gets parsed to what it could and leaves the rest behind, resulting in just `anything;` being left.
<p>Publish Date: 2018-05-31
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10531>CVE-2016-10531</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nodesecurity.io/advisories/101">https://nodesecurity.io/advisories/101</a></p>
<p>Release Date: 2016-04-18</p>
<p>Fix Resolution: Update to version 0.3.6 or later.</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"marked","packageVersion":"0.3.19","packageFilePaths":["/docs/node_modules/marked/www/demo.html","/docs-aspnet/node_modules/marked/www/demo.html"],"isTransitiveDependency":false,"dependencyTree":"marked:0.3.19","isMinimumFixVersionAvailable":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2016-10531","vulnerabilityDetails":"marked is an application that is meant to parse and compile markdown. Due to the way that marked 0.3.5 and earlier parses input, specifically HTML entities, it\u0027s possible to bypass marked\u0027s content injection protection (`sanitize: true`) to inject a `javascript:` URL. This flaw exists because `\u0026#xNNanything;` gets parsed to what it could and leaves the rest behind, resulting in just `anything;` being left.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10531","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | True | CVE-2016-10531 (Medium) detected in marked-0.3.19.js - ## CVE-2016-10531 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>marked-0.3.19.js</b></p></summary>
<p>A markdown parser built for speed</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/marked/0.3.19/marked.js">https://cdnjs.cloudflare.com/ajax/libs/marked/0.3.19/marked.js</a></p>
<p>Path to dependency file: kendo-ui-core/docs/node_modules/marked/www/demo.html</p>
<p>Path to vulnerable library: kendo-ui-core/docs/node_modules/marked/www/../lib/marked.js,kendo-ui-core/docs-aspnet/node_modules/marked/www/../lib/marked.js</p>
<p>
Dependency Hierarchy:
- :x: **marked-0.3.19.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/rsoreq/kendo-ui-core/commit/62afbcdf79c4c7052417ecc86eb31bd6bc04e1ad">62afbcdf79c4c7052417ecc86eb31bd6bc04e1ad</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
marked is an application that is meant to parse and compile markdown. Due to the way that marked 0.3.5 and earlier parses input, specifically HTML entities, it's possible to bypass marked's content injection protection (`sanitize: true`) to inject a `javascript:` URL. This flaw exists because `&#xNNanything;` gets parsed to what it could and leaves the rest behind, resulting in just `anything;` being left.
<p>Publish Date: 2018-05-31
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10531>CVE-2016-10531</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nodesecurity.io/advisories/101">https://nodesecurity.io/advisories/101</a></p>
<p>Release Date: 2016-04-18</p>
<p>Fix Resolution: Update to version 0.3.6 or later.</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"marked","packageVersion":"0.3.19","packageFilePaths":["/docs/node_modules/marked/www/demo.html","/docs-aspnet/node_modules/marked/www/demo.html"],"isTransitiveDependency":false,"dependencyTree":"marked:0.3.19","isMinimumFixVersionAvailable":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2016-10531","vulnerabilityDetails":"marked is an application that is meant to parse and compile markdown. Due to the way that marked 0.3.5 and earlier parses input, specifically HTML entities, it\u0027s possible to bypass marked\u0027s content injection protection (`sanitize: true`) to inject a `javascript:` URL. This flaw exists because `\u0026#xNNanything;` gets parsed to what it could and leaves the rest behind, resulting in just `anything;` being left.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-10531","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | non_defect | cve medium detected in marked js cve medium severity vulnerability vulnerable library marked js a markdown parser built for speed library home page a href path to dependency file kendo ui core docs node modules marked www demo html path to vulnerable library kendo ui core docs node modules marked www lib marked js kendo ui core docs aspnet node modules marked www lib marked js dependency hierarchy x marked js vulnerable library found in head commit a href found in base branch master vulnerability details marked is an application that is meant to parse and compile markdown due to the way that marked and earlier parses input specifically html entities it s possible to bypass marked s content injection protection sanitize true to inject a javascript url this flaw exists because xnnanything gets parsed to what it could and leaves the rest behind resulting in just anything being left publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution update to version or later isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree marked isminimumfixversionavailable false basebranches vulnerabilityidentifier cve vulnerabilitydetails marked is an application that is meant to parse and compile markdown due to the way that marked and earlier parses input specifically html entities it possible to bypass marked content injection protection sanitize true to inject a javascript url this flaw exists because xnnanything gets parsed to what it could and leaves the rest behind resulting in just anything being left vulnerabilityurl | 0 |
56,787 | 15,370,330,800 | IssuesEvent | 2021-03-02 08:40:05 | MethodicalAcceleratorDesign/MAD-X | https://api.github.com/repos/MethodicalAcceleratorDesign/MAD-X | closed | Tilt + misalignment behaves strange in MAD-X PTC | PTC bug defect | It looks to be like the misalignment is applied in the tilted frame and is therefore moving the magnet in the wrong direction. I attach an example to illustrate:
FASER : SEQUENCE, L = 10;
b1:sbend, angle=138/180.0*pi, HGAP=0.0258,e1=0/180.0*pi, e2= 0/180.0*pi,FINT=0.5 ,TILT=1.0*pi, at=5,l=1;
endsequence;
beam;
use, sequence=faser;
eoption, add=true;
select, flag=error, clear=true;
select, flag=error, class=b1;
ealign, dx= 0.001, dy= 0.00, ds=0.00,dphi=0.00, dtheta=0.00, dpsi=0.00;
select, flag=twiss, column=name,s,x,px,y, py;
twiss, betx=1, bety=1, x=-0.000, file="mytwiss.dat";
ptc_create_universe;
ptc_create_layout, model=2, method=6,nst=10, exact;
ptc_align;
ptc_start, x=0.000;
ptc_observe, place=b1;
ptc_track, icase=5, element_by_element, dump, file="track.simple", ffile=1, onetable, maxaper={100.0, 100.0, 100.0, 100.0, 100.0, 100.0};
ptc_twiss, file="twiss.ptc.simple",betx=1,bety=1,betz=1;
ptc_track_end;
ptc_end; | 1.0 | Tilt + misalignment behaves strange in MAD-X PTC - It looks to be like the misalignment is applied in the tilted frame and is therefore moving the magnet in the wrong direction. I attach an example to illustrate:
FASER : SEQUENCE, L = 10;
b1:sbend, angle=138/180.0*pi, HGAP=0.0258,e1=0/180.0*pi, e2= 0/180.0*pi,FINT=0.5 ,TILT=1.0*pi, at=5,l=1;
endsequence;
beam;
use, sequence=faser;
eoption, add=true;
select, flag=error, clear=true;
select, flag=error, class=b1;
ealign, dx= 0.001, dy= 0.00, ds=0.00,dphi=0.00, dtheta=0.00, dpsi=0.00;
select, flag=twiss, column=name,s,x,px,y, py;
twiss, betx=1, bety=1, x=-0.000, file="mytwiss.dat";
ptc_create_universe;
ptc_create_layout, model=2, method=6,nst=10, exact;
ptc_align;
ptc_start, x=0.000;
ptc_observe, place=b1;
ptc_track, icase=5, element_by_element, dump, file="track.simple", ffile=1, onetable, maxaper={100.0, 100.0, 100.0, 100.0, 100.0, 100.0};
ptc_twiss, file="twiss.ptc.simple",betx=1,bety=1,betz=1;
ptc_track_end;
ptc_end; | defect | tilt misalignment behaves strange in mad x ptc it looks to be like the misalignment is applied in the tilted frame and is therefore moving the magnet in the wrong direction i attach an example to illustrate faser sequence l sbend angle pi hgap pi pi fint tilt pi at l endsequence beam use sequence faser eoption add true select flag error clear true select flag error class ealign dx dy ds dphi dtheta dpsi select flag twiss column name s x px y py twiss betx bety x file mytwiss dat ptc create universe ptc create layout model method nst exact ptc align ptc start x ptc observe place ptc track icase element by element dump file track simple ffile onetable maxaper ptc twiss file twiss ptc simple betx bety betz ptc track end ptc end | 1 |
10,115 | 7,918,964,508 | IssuesEvent | 2018-07-04 14:58:25 | maxdobeck/scheduler-frontend | https://api.github.com/repos/maxdobeck/scheduler-frontend | opened | Check for Good Session In Route Guard | bug security | User should be logged out or forced to log in again when session ends. | True | Check for Good Session In Route Guard - User should be logged out or forced to log in again when session ends. | non_defect | check for good session in route guard user should be logged out or forced to log in again when session ends | 0 |
39,678 | 9,612,973,604 | IssuesEvent | 2019-05-13 10:09:01 | contao/contao | https://api.github.com/repos/contao/contao | closed | Symlinks neu erstellen | defect | <a href="https://github.com/jommbee"><img src="https://avatars3.githubusercontent.com/u/4679889?v=4" align="left" width="42" height="42"></img></a> [Comment](https://github.com/contao/contao/issues/299#issuecomment-487871058) by @jommbee
April 30th, 2019, 10:59 GMT
Heute in der 4.4.39 nach Serverumzug gleicher Fehler.
Nach manuellem Fix in der Zeile:
https://github.com/contao/contao/blob/4.4/core-bundle/src/Resources/contao/library/Contao/Automator.php#L400
geht alles wunderbar.
| 1.0 | Symlinks neu erstellen - <a href="https://github.com/jommbee"><img src="https://avatars3.githubusercontent.com/u/4679889?v=4" align="left" width="42" height="42"></img></a> [Comment](https://github.com/contao/contao/issues/299#issuecomment-487871058) by @jommbee
April 30th, 2019, 10:59 GMT
Heute in der 4.4.39 nach Serverumzug gleicher Fehler.
Nach manuellem Fix in der Zeile:
https://github.com/contao/contao/blob/4.4/core-bundle/src/Resources/contao/library/Contao/Automator.php#L400
geht alles wunderbar.
| defect | symlinks neu erstellen by jommbee april gmt heute in der nach serverumzug gleicher fehler nach manuellem fix in der zeile geht alles wunderbar | 1 |
240,180 | 26,254,331,125 | IssuesEvent | 2023-01-05 22:33:20 | TreyM-WSS/terra-clinical | https://api.github.com/repos/TreyM-WSS/terra-clinical | opened | CVE-2021-23362 (Medium) detected in hosted-git-info-2.8.8.tgz | security vulnerability | ## CVE-2021-23362 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hosted-git-info-2.8.8.tgz</b></p></summary>
<p>Provides metadata and conversions from repository urls for Github, Bitbucket and Gitlab</p>
<p>Library home page: <a href="https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz">https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/hosted-git-info/package.json</p>
<p>
Dependency Hierarchy:
- check-installed-dependencies-1.0.1.tgz (Root Library)
- meow-3.7.0.tgz
- normalize-package-data-2.5.0.tgz
- :x: **hosted-git-info-2.8.8.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/TreyM-WSS/terra-clinical/commit/0940364474cdf2812b34bb24824139d7e3f8060b">0940364474cdf2812b34bb24824139d7e3f8060b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity.
<p>Publish Date: 2021-03-23
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-23362>CVE-2021-23362</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-43f8-2h32-f4cj">https://github.com/advisories/GHSA-43f8-2h32-f4cj</a></p>
<p>Release Date: 2021-03-23</p>
<p>Fix Resolution: hosted-git-info - 2.8.9,3.0.8</p>
</p>
</details>
<p></p>
| True | CVE-2021-23362 (Medium) detected in hosted-git-info-2.8.8.tgz - ## CVE-2021-23362 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>hosted-git-info-2.8.8.tgz</b></p></summary>
<p>Provides metadata and conversions from repository urls for Github, Bitbucket and Gitlab</p>
<p>Library home page: <a href="https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz">https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/hosted-git-info/package.json</p>
<p>
Dependency Hierarchy:
- check-installed-dependencies-1.0.1.tgz (Root Library)
- meow-3.7.0.tgz
- normalize-package-data-2.5.0.tgz
- :x: **hosted-git-info-2.8.8.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/TreyM-WSS/terra-clinical/commit/0940364474cdf2812b34bb24824139d7e3f8060b">0940364474cdf2812b34bb24824139d7e3f8060b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity.
<p>Publish Date: 2021-03-23
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-23362>CVE-2021-23362</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-43f8-2h32-f4cj">https://github.com/advisories/GHSA-43f8-2h32-f4cj</a></p>
<p>Release Date: 2021-03-23</p>
<p>Fix Resolution: hosted-git-info - 2.8.9,3.0.8</p>
</p>
</details>
<p></p>
| non_defect | cve medium detected in hosted git info tgz cve medium severity vulnerability vulnerable library hosted git info tgz provides metadata and conversions from repository urls for github bitbucket and gitlab library home page a href path to dependency file package json path to vulnerable library node modules hosted git info package json dependency hierarchy check installed dependencies tgz root library meow tgz normalize package data tgz x hosted git info tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package hosted git info before are vulnerable to regular expression denial of service redos via regular expression shortcutmatch in the fromurl function in index js the affected regular expression exhibits polynomial worst case time complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution hosted git info | 0 |
15,626 | 2,866,060,794 | IssuesEvent | 2015-06-05 03:21:12 | chaosben/theunknownones | https://api.github.com/repos/chaosben/theunknownones | closed | Fermeture of Google Code | auto-migrated Priority-Medium Type-Defect | ```
Google Code will soon close. Have you thinked of another place to migrate, like
GitHub (Using the "Export to GitHub" button)?
```
Original issue reported on code.google.com by `charlesm...@gmail.com` on 13 May 2015 at 10:15 | 1.0 | Fermeture of Google Code - ```
Google Code will soon close. Have you thinked of another place to migrate, like
GitHub (Using the "Export to GitHub" button)?
```
Original issue reported on code.google.com by `charlesm...@gmail.com` on 13 May 2015 at 10:15 | defect | fermeture of google code google code will soon close have you thinked of another place to migrate like github using the export to github button original issue reported on code google com by charlesm gmail com on may at | 1 |
74,821 | 25,345,831,442 | IssuesEvent | 2022-11-19 07:09:06 | openzfs/zfs | https://api.github.com/repos/openzfs/zfs | opened | fedora 37 upgrade fails when importing gpg keys | Type: Defect | <!-- Please fill out the following template, which will help other contributors address your issue. -->
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please check our issue tracker before opening a new issue.
Additional valuable information can be found in the OpenZFS documentation
and mailing list archives.
Please fill in as much of the template as possible.
-->
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
-->
Distribution Name | fedora
Distribution Version | 36
Kernel Version | 6.0.6-200.fc36
Architecture | x86-64
OpenZFS Version | zfs-2.1.6-1
<!--
Command to find OpenZFS version:
zfs version
Commands to find kernel version:
uname -r # Linux
freebsd-version -r # FreeBSD
-->
### Describe the problem you're observing
system-upgrade to --releasever=37 fails when dnf attempts to import gpg keys for zfs fc27 rpms
### Describe how to reproduce the problem
sudo dnf -y system-upgrade download --allowerasing --refresh --releasever=37
dnf downloads zfs rpms from fc37 repo but fails when attempting to import gpg keys.
I didn't see this problem when doing a system-upgrade to fc36.
### Include any warning/errors/backtraces from the system logs
<!--
*IMPORTANT* - Please mark logs and text output from terminal commands
or else Github will not display them correctly.
An example is provided below.
Example:
-->
```
Fedora 37 - x86_64 1.6 MB/s | 1.6 kB 00:00
Importing GPG key 0x5323552A:
Userid : "Fedora (37) <fedora-37-primary@fedoraproject.org>"
Fingerprint: ACB5 EE4E 831C 74BB 7C16 8D27 F55A D3FB 5323 552A
From : /etc/pki/rpm-gpg/RPM-GPG-KEY-fedora-37-x86_64
Key imported successfully
ZFS on Linux for Fedora 37 1.6 MB/s | 1.7 kB 00:00
GPG key at file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux (0xF14AB620) is already installed
The GPG keys listed for the "ZFS on Linux for Fedora 37" repository are already installed but they
are not correct for this package.
Check that the correct key URLs are configured for this repository..
Failing package is: libnvpair3-2.1.6-2.fc37.x86_64
GPG Keys are configured as: file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux
Public key for libuutil3-2.1.6-2.fc37.x86_64.rpm is not installed.
Failing package is: libuutil3-2.1.6-2.fc37.x86_64
GPG Keys are configured as: file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux
Public key for libzfs5-2.1.6-2.fc37.x86_64.rpm is not installed.
Failing package is: libzfs5-2.1.6-2.fc37.x86_64
GPG Keys are configured as: file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux
Public key for libzpool5-2.1.6-2.fc37.x86_64.rpm is not installed.
Failing package is: libzpool5-2.1.6-2.fc37.x86_64
GPG Keys are configured as: file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux
Public key for zfs-2.1.6-2.fc37.x86_64.rpm is not installed.
Failing package is: zfs-2.1.6-2.fc37.x86_64
GPG Keys are configured as: file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux
Public key for zfs-dkms-2.1.6-2.fc37.noarch.rpm is not installed.
Failing package is: zfs-dkms-2.1.6-2.fc37.noarch
GPG Keys are configured as: file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux
Public key for zfs-dracut-2.1.6-2.fc37.noarch.rpm is not installed.
Failing package is: zfs-dracut-2.1.6-2.fc37.noarch
GPG Keys are configured as: file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux
The downloaded packages were saved in cache until the next successful transaction.
You can remove cached packages by executing 'dnf clean packages'.
Error: GPG check FAILED
```
file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux | 1.0 | fedora 37 upgrade fails when importing gpg keys - <!-- Please fill out the following template, which will help other contributors address your issue. -->
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please check our issue tracker before opening a new issue.
Additional valuable information can be found in the OpenZFS documentation
and mailing list archives.
Please fill in as much of the template as possible.
-->
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
-->
Distribution Name | fedora
Distribution Version | 36
Kernel Version | 6.0.6-200.fc36
Architecture | x86-64
OpenZFS Version | zfs-2.1.6-1
<!--
Command to find OpenZFS version:
zfs version
Commands to find kernel version:
uname -r # Linux
freebsd-version -r # FreeBSD
-->
### Describe the problem you're observing
system-upgrade to --releasever=37 fails when dnf attempts to import gpg keys for zfs fc27 rpms
### Describe how to reproduce the problem
sudo dnf -y system-upgrade download --allowerasing --refresh --releasever=37
dnf downloads zfs rpms from fc37 repo but fails when attempting to import gpg keys.
I didn't see this problem when doing a system-upgrade to fc36.
### Include any warning/errors/backtraces from the system logs
<!--
*IMPORTANT* - Please mark logs and text output from terminal commands
or else Github will not display them correctly.
An example is provided below.
Example:
-->
```
Fedora 37 - x86_64 1.6 MB/s | 1.6 kB 00:00
Importing GPG key 0x5323552A:
Userid : "Fedora (37) <fedora-37-primary@fedoraproject.org>"
Fingerprint: ACB5 EE4E 831C 74BB 7C16 8D27 F55A D3FB 5323 552A
From : /etc/pki/rpm-gpg/RPM-GPG-KEY-fedora-37-x86_64
Key imported successfully
ZFS on Linux for Fedora 37 1.6 MB/s | 1.7 kB 00:00
GPG key at file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux (0xF14AB620) is already installed
The GPG keys listed for the "ZFS on Linux for Fedora 37" repository are already installed but they
are not correct for this package.
Check that the correct key URLs are configured for this repository..
Failing package is: libnvpair3-2.1.6-2.fc37.x86_64
GPG Keys are configured as: file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux
Public key for libuutil3-2.1.6-2.fc37.x86_64.rpm is not installed.
Failing package is: libuutil3-2.1.6-2.fc37.x86_64
GPG Keys are configured as: file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux
Public key for libzfs5-2.1.6-2.fc37.x86_64.rpm is not installed.
Failing package is: libzfs5-2.1.6-2.fc37.x86_64
GPG Keys are configured as: file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux
Public key for libzpool5-2.1.6-2.fc37.x86_64.rpm is not installed.
Failing package is: libzpool5-2.1.6-2.fc37.x86_64
GPG Keys are configured as: file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux
Public key for zfs-2.1.6-2.fc37.x86_64.rpm is not installed.
Failing package is: zfs-2.1.6-2.fc37.x86_64
GPG Keys are configured as: file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux
Public key for zfs-dkms-2.1.6-2.fc37.noarch.rpm is not installed.
Failing package is: zfs-dkms-2.1.6-2.fc37.noarch
GPG Keys are configured as: file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux
Public key for zfs-dracut-2.1.6-2.fc37.noarch.rpm is not installed.
Failing package is: zfs-dracut-2.1.6-2.fc37.noarch
GPG Keys are configured as: file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux
The downloaded packages were saved in cache until the next successful transaction.
You can remove cached packages by executing 'dnf clean packages'.
Error: GPG check FAILED
```
file:///etc/pki/rpm-gpg/RPM-GPG-KEY-zfsonlinux | defect | fedora upgrade fails when importing gpg keys thank you for reporting an issue important please check our issue tracker before opening a new issue additional valuable information can be found in the openzfs documentation and mailing list archives please fill in as much of the template as possible system information type version name distribution name fedora distribution version kernel version architecture openzfs version zfs command to find openzfs version zfs version commands to find kernel version uname r linux freebsd version r freebsd describe the problem you re observing system upgrade to releasever fails when dnf attempts to import gpg keys for zfs rpms describe how to reproduce the problem sudo dnf y system upgrade download allowerasing refresh releasever dnf downloads zfs rpms from repo but fails when attempting to import gpg keys i didn t see this problem when doing a system upgrade to include any warning errors backtraces from the system logs important please mark logs and text output from terminal commands or else github will not display them correctly an example is provided below example fedora mb s kb importing gpg key userid fedora fingerprint from etc pki rpm gpg rpm gpg key fedora key imported successfully zfs on linux for fedora mb s kb gpg key at file etc pki rpm gpg rpm gpg key zfsonlinux is already installed the gpg keys listed for the zfs on linux for fedora repository are already installed but they are not correct for this package check that the correct key urls are configured for this repository failing package is gpg keys are configured as file etc pki rpm gpg rpm gpg key zfsonlinux public key for rpm is not installed failing package is gpg keys are configured as file etc pki rpm gpg rpm gpg key zfsonlinux public key for rpm is not installed failing package is gpg keys are configured as file etc pki rpm gpg rpm gpg key zfsonlinux public key for rpm is not installed failing package is gpg keys are configured as file etc pki rpm gpg rpm gpg key zfsonlinux public key for zfs rpm is not installed failing package is zfs gpg keys are configured as file etc pki rpm gpg rpm gpg key zfsonlinux public key for zfs dkms noarch rpm is not installed failing package is zfs dkms noarch gpg keys are configured as file etc pki rpm gpg rpm gpg key zfsonlinux public key for zfs dracut noarch rpm is not installed failing package is zfs dracut noarch gpg keys are configured as file etc pki rpm gpg rpm gpg key zfsonlinux the downloaded packages were saved in cache until the next successful transaction you can remove cached packages by executing dnf clean packages error gpg check failed file etc pki rpm gpg rpm gpg key zfsonlinux | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.