Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 2 665 | labels stringlengths 4 554 | body stringlengths 3 235k | index stringclasses 6 values | text_combine stringlengths 96 235k | label stringclasses 2 values | text stringlengths 96 196k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
19,807 | 13,464,926,086 | IssuesEvent | 2020-09-09 19:59:54 | nanovms/nanos | https://api.github.com/repos/nanovms/nanos | closed | openstack rollup | infrastructure openstack | not sure what all is necessary to support - it uses kvm underneath w/virtio-net && virtio-block (and also virtio-rng)

however currently it is just rebooting (uploaded as a raw disk - maybe there are different options to specify when provisioning
this was done using devstack under fusion so it could be the nested virt as well - idk
https://docs.openstack.org/devstack/latest/
| 1.0 | openstack rollup - not sure what all is necessary to support - it uses kvm underneath w/virtio-net && virtio-block (and also virtio-rng)

however currently it is just rebooting (uploaded as a raw disk - maybe there are different options to specify when provisioning
this was done using devstack under fusion so it could be the nested virt as well - idk
https://docs.openstack.org/devstack/latest/
| infrastructure | openstack rollup not sure what all is necessary to support it uses kvm underneath w virtio net virtio block and also virtio rng however currently it is just rebooting uploaded as a raw disk maybe there are different options to specify when provisioning this was done using devstack under fusion so it could be the nested virt as well idk | 1 |
7,754 | 7,083,501,944 | IssuesEvent | 2018-01-11 00:42:50 | dotnet/roslyn | https://api.github.com/repos/dotnet/roslyn | closed | Update LangVersion for Roslyn VB projects to 15.3 or 15.5 | Area-Infrastructure | When I tried, the Ubuntu CI builds failed with: `error BC2014: the value '15.3' is invalid for option 'langversion' `
This suggests that the .NET Core build may not be using the package specified by `MicrosoftNetCompilersnetcoreVersion`.
FYI @agocke | 1.0 | Update LangVersion for Roslyn VB projects to 15.3 or 15.5 - When I tried, the Ubuntu CI builds failed with: `error BC2014: the value '15.3' is invalid for option 'langversion' `
This suggests that the .NET Core build may not be using the package specified by `MicrosoftNetCompilersnetcoreVersion`.
FYI @agocke | infrastructure | update langversion for roslyn vb projects to or when i tried the ubuntu ci builds failed with error the value is invalid for option langversion this suggests that the net core build may not be using the package specified by microsoftnetcompilersnetcoreversion fyi agocke | 1 |
56,264 | 31,834,402,671 | IssuesEvent | 2023-09-14 12:42:40 | IGCIT/Intel-GPU-Community-Issue-Tracker-IGCIT | https://api.github.com/repos/IGCIT/Intel-GPU-Community-Issue-Tracker-IGCIT | closed | Valorant Shuttering and no Multithereading | Info: Needs Replication Type: Bad Performance | ### Checklist [README]
- [X] Device is not undervolted nor overclocked
- [X] Device is using the [latest drivers](https://github.com/IGCIT/Intel-GPU-Community-Issue-Tracker-IGCIT/wiki/Intel-driver-updates-and-changelogs)
- [X] Application is not cracked, modded and use the latest patch
### Application [Required]
Valorant
### Processor / Processor Number [Required]
Intel i5 13600K
### Graphic Card [Required]
Intel Arc A770 Limited Edition
### GPU Driver Version [Required]
- [X] 31.0.101.4676
- [ ] 31.0.101.4672
- [ ] 31.0.101.4669
- [ ] 31.0.101.4644
- [ ] 31.0.101.4578
- [ ] 31.0.101.4577
- [ ] 31.0.101.4576
- [ ] 31.0.101.4575
- [ ] 31.0.101.4514
- [ ] 31.0.101.4502
- [ ] 31.0.101.4499
- [ ] 31.0.101.4382
- [ ] 31.0.101.4369
- [ ] 31.0.101.4335
- [ ] 31.0.101.4314
- [ ] 31.0.101.4311
- [ ] 31.0.101.4257
- [ ] 31.0.101.4255
- [ ] 31.0.101.4148
- [ ] 31.0.101.4146
- [ ] 31.0.101.4125
- [ ] 31.0.101.4123
- [ ] 31.0.101.4091
- [ ] 31.0.101.4090
- [ ] 31.0.101.4034
### Rendering API [Required]
- [ ] Vulkan
- [ ] OpenGL
- [ ] DirectX12
- [X] DirectX11
- [ ] DirectX10
- [ ] DirectX9
- [ ] Not applicable
### Windows Build Number [Required]
- [X] Windows 11 22H2
- [ ] Windows 11 21H2
- [ ] Windows 10 22H2
- [ ] Windows 10 21H2
- [ ] Other (Please specify)
### Other Windows build number
_No response_
### Intel System Support Utility report
[ssulog.txt](https://github.com/IGCIT/Intel-GPU-Community-Issue-Tracker-IGCIT/files/12565294/ssulog.txt)
### Description and steps to reproduce [Required]
Just install the Intel Arc A770 Graphicscard into Ur pc and install Valorant and go for a round. for the Multithreading option go to Settings and then go to Video and then u can see the Multithreading Option is missing.
### Device / Platform
_No response_
### Crash dumps [Required, if applicable]
_No response_
### Application / Windows logs
_No response_ | True | Valorant Shuttering and no Multithereading - ### Checklist [README]
- [X] Device is not undervolted nor overclocked
- [X] Device is using the [latest drivers](https://github.com/IGCIT/Intel-GPU-Community-Issue-Tracker-IGCIT/wiki/Intel-driver-updates-and-changelogs)
- [X] Application is not cracked, modded and use the latest patch
### Application [Required]
Valorant
### Processor / Processor Number [Required]
Intel i5 13600K
### Graphic Card [Required]
Intel Arc A770 Limited Edition
### GPU Driver Version [Required]
- [X] 31.0.101.4676
- [ ] 31.0.101.4672
- [ ] 31.0.101.4669
- [ ] 31.0.101.4644
- [ ] 31.0.101.4578
- [ ] 31.0.101.4577
- [ ] 31.0.101.4576
- [ ] 31.0.101.4575
- [ ] 31.0.101.4514
- [ ] 31.0.101.4502
- [ ] 31.0.101.4499
- [ ] 31.0.101.4382
- [ ] 31.0.101.4369
- [ ] 31.0.101.4335
- [ ] 31.0.101.4314
- [ ] 31.0.101.4311
- [ ] 31.0.101.4257
- [ ] 31.0.101.4255
- [ ] 31.0.101.4148
- [ ] 31.0.101.4146
- [ ] 31.0.101.4125
- [ ] 31.0.101.4123
- [ ] 31.0.101.4091
- [ ] 31.0.101.4090
- [ ] 31.0.101.4034
### Rendering API [Required]
- [ ] Vulkan
- [ ] OpenGL
- [ ] DirectX12
- [X] DirectX11
- [ ] DirectX10
- [ ] DirectX9
- [ ] Not applicable
### Windows Build Number [Required]
- [X] Windows 11 22H2
- [ ] Windows 11 21H2
- [ ] Windows 10 22H2
- [ ] Windows 10 21H2
- [ ] Other (Please specify)
### Other Windows build number
_No response_
### Intel System Support Utility report
[ssulog.txt](https://github.com/IGCIT/Intel-GPU-Community-Issue-Tracker-IGCIT/files/12565294/ssulog.txt)
### Description and steps to reproduce [Required]
Just install the Intel Arc A770 Graphicscard into Ur pc and install Valorant and go for a round. for the Multithreading option go to Settings and then go to Video and then u can see the Multithreading Option is missing.
### Device / Platform
_No response_
### Crash dumps [Required, if applicable]
_No response_
### Application / Windows logs
_No response_ | non_infrastructure | valorant shuttering and no multithereading checklist device is not undervolted nor overclocked device is using the application is not cracked modded and use the latest patch application valorant processor processor number intel graphic card intel arc limited edition gpu driver version rendering api vulkan opengl not applicable windows build number windows windows windows windows other please specify other windows build number no response intel system support utility report description and steps to reproduce just install the intel arc graphicscard into ur pc and install valorant and go for a round for the multithreading option go to settings and then go to video and then u can see the multithreading option is missing device platform no response crash dumps no response application windows logs no response | 0 |
1,238 | 3,089,244,652 | IssuesEvent | 2015-08-25 20:33:59 | dotnet/coreclr | https://api.github.com/repos/dotnet/coreclr | closed | Update coreclr tests to use newer json package format | 2 - In Progress infrastructure | The tests should be updated so that the use the newer package format. | 1.0 | Update coreclr tests to use newer json package format - The tests should be updated so that the use the newer package format. | infrastructure | update coreclr tests to use newer json package format the tests should be updated so that the use the newer package format | 1 |
19,489 | 13,255,766,860 | IssuesEvent | 2020-08-20 11:32:16 | nhsx/covid-chest-imaging-database | https://api.github.com/repos/nhsx/covid-chest-imaging-database | closed | Run the pipeline as a Fargate task | enhancement infrastructure | ### Describe the feature request
Package up and run the pipeline as a [Fargate task](https://docs.aws.amazon.com/AmazonECS/latest/developerguide/AWS_Fargate.html), independently from the platform.
### Affected product area (please put an X in all the brackets that apply)
[ ] Documentation
[ ] Data access
[ ] Infrastructure
[x] Loader pipeline
[ ] Other (please describe)
### Additional context
Requires roughly:
* ECS container
* Task definition
* Recurring schedule
* relevant infrastructure (ECR, isolated vpc, security groups, s3 endpoint, ...) defined in template
| 1.0 | Run the pipeline as a Fargate task - ### Describe the feature request
Package up and run the pipeline as a [Fargate task](https://docs.aws.amazon.com/AmazonECS/latest/developerguide/AWS_Fargate.html), independently from the platform.
### Affected product area (please put an X in all the brackets that apply)
[ ] Documentation
[ ] Data access
[ ] Infrastructure
[x] Loader pipeline
[ ] Other (please describe)
### Additional context
Requires roughly:
* ECS container
* Task definition
* Recurring schedule
* relevant infrastructure (ECR, isolated vpc, security groups, s3 endpoint, ...) defined in template
| infrastructure | run the pipeline as a fargate task describe the feature request package up and run the pipeline as a independently from the platform affected product area please put an x in all the brackets that apply documentation data access infrastructure loader pipeline other please describe additional context requires roughly ecs container task definition recurring schedule relevant infrastructure ecr isolated vpc security groups endpoint defined in template | 1 |
9,633 | 8,058,434,632 | IssuesEvent | 2018-08-02 18:27:44 | dotnet/coreclr | https://api.github.com/repos/dotnet/coreclr | closed | Alpine tests are all broken | area-Infrastructure test bug | https://mc.dot.net/#/product/netcore/30/source/official~2Fcoreclr~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180727.01
It seems some script issue:
```
2018-06-26 12:20:06,991: INFO: proc(54): run_and_log_output: Output: /home/helixbot/dotnetbuild/work/6b9941f0-db03-4533-a536-11716462b85f/Work/1225557e-52ac-41f7-8b62-8428a2169d36/Unzip/runtests.sh: line 13: /home/helixbot/dotnetbuild/work/6b9941f0-db03-4533-a536-11716462b85f/Payload/corerun: not found
2018-06-26 12:20:06,991: INFO: proc(54): run_and_log_output: Output: Finished running tests. Exit code = 127
```
Looking back, they have been broken for at least a month so we may not be getting any Alpine coverage for that time at least in official runs.
@RussKeldorph | 1.0 | Alpine tests are all broken - https://mc.dot.net/#/product/netcore/30/source/official~2Fcoreclr~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180727.01
It seems some script issue:
```
2018-06-26 12:20:06,991: INFO: proc(54): run_and_log_output: Output: /home/helixbot/dotnetbuild/work/6b9941f0-db03-4533-a536-11716462b85f/Work/1225557e-52ac-41f7-8b62-8428a2169d36/Unzip/runtests.sh: line 13: /home/helixbot/dotnetbuild/work/6b9941f0-db03-4533-a536-11716462b85f/Payload/corerun: not found
2018-06-26 12:20:06,991: INFO: proc(54): run_and_log_output: Output: Finished running tests. Exit code = 127
```
Looking back, they have been broken for at least a month so we may not be getting any Alpine coverage for that time at least in official runs.
@RussKeldorph | infrastructure | alpine tests are all broken it seems some script issue info proc run and log output output home helixbot dotnetbuild work work unzip runtests sh line home helixbot dotnetbuild work payload corerun not found info proc run and log output output finished running tests exit code looking back they have been broken for at least a month so we may not be getting any alpine coverage for that time at least in official runs russkeldorph | 1 |
15,605 | 11,610,980,274 | IssuesEvent | 2020-02-26 05:01:34 | pyinstaller/pyinstaller | https://api.github.com/repos/pyinstaller/pyinstaller | closed | Switch from Github Services to webhooks | area:project infrastructure help wanted | [Github services](https://github.com/pyinstaller/pyinstaller/settings/installations) are deprecated. We need to migrate to webhooks or GitHub Apps.
Currently we have these services in place (tick of what has been migrated)
- [ ] Email
- [ ] Landscape
- [X] ReadTheDocs
- [ ] Travis CI | 1.0 | Switch from Github Services to webhooks - [Github services](https://github.com/pyinstaller/pyinstaller/settings/installations) are deprecated. We need to migrate to webhooks or GitHub Apps.
Currently we have these services in place (tick of what has been migrated)
- [ ] Email
- [ ] Landscape
- [X] ReadTheDocs
- [ ] Travis CI | infrastructure | switch from github services to webhooks are deprecated we need to migrate to webhooks or github apps currently we have these services in place tick of what has been migrated email landscape readthedocs travis ci | 1 |
27,470 | 21,774,125,952 | IssuesEvent | 2022-05-13 12:10:40 | ably/ably-flutter | https://api.github.com/repos/ably/ably-flutter | opened | Fix consistency of test dependencies | test-infrastructure | I've noticed that both `flutter_test` and `test` are used in the project. Usually, it's only required to use one of these packages, (since `flutter_test` uses `test` as it's core) and if they're both used at the same time, they may cause conflicts. It would be good to do a dependency cleanup for tests and leave only one of those | 1.0 | Fix consistency of test dependencies - I've noticed that both `flutter_test` and `test` are used in the project. Usually, it's only required to use one of these packages, (since `flutter_test` uses `test` as it's core) and if they're both used at the same time, they may cause conflicts. It would be good to do a dependency cleanup for tests and leave only one of those | infrastructure | fix consistency of test dependencies i ve noticed that both flutter test and test are used in the project usually it s only required to use one of these packages since flutter test uses test as it s core and if they re both used at the same time they may cause conflicts it would be good to do a dependency cleanup for tests and leave only one of those | 1 |
36,852 | 15,085,206,466 | IssuesEvent | 2021-02-05 18:18:26 | retaildevcrews/ngsa | https://api.github.com/repos/retaildevcrews/ngsa | closed | add log message | DataService Observability Observe | - when we get a 429 response and serve from cache, we should log the 429
- currently, it shows as a 200 (which is correct, but loses fidelity) | 1.0 | add log message - - when we get a 429 response and serve from cache, we should log the 429
- currently, it shows as a 200 (which is correct, but loses fidelity) | non_infrastructure | add log message when we get a response and serve from cache we should log the currently it shows as a which is correct but loses fidelity | 0 |
14,124 | 10,619,145,863 | IssuesEvent | 2019-10-13 11:05:50 | dotnet/corefx | https://api.github.com/repos/dotnet/corefx | closed | CoreFx assemblies lack the TargetFrameworkAttribute | area-Infrastructure bug | I notice this when looking at https://github.com/dotnet/corefx/issues/26456. It appears to still be the case. Filing here, though the right fix may be in buildtools. | 1.0 | CoreFx assemblies lack the TargetFrameworkAttribute - I notice this when looking at https://github.com/dotnet/corefx/issues/26456. It appears to still be the case. Filing here, though the right fix may be in buildtools. | infrastructure | corefx assemblies lack the targetframeworkattribute i notice this when looking at it appears to still be the case filing here though the right fix may be in buildtools | 1 |
6,566 | 6,515,885,875 | IssuesEvent | 2017-08-26 22:08:33 | SatelliteQE/robottelo-ci | https://api.github.com/repos/SatelliteQE/robottelo-ci | closed | Vlan Sat6major_el7 missing on vmware , depending test's will fail | Infrastructure | Vlan is needed to provision boxes in specific subnet, for our automation especially for 6.3 , we use `sat6major_el7` bridge , which is missing in vmware setup | 1.0 | Vlan Sat6major_el7 missing on vmware , depending test's will fail - Vlan is needed to provision boxes in specific subnet, for our automation especially for 6.3 , we use `sat6major_el7` bridge , which is missing in vmware setup | infrastructure | vlan missing on vmware depending test s will fail vlan is needed to provision boxes in specific subnet for our automation especially for we use bridge which is missing in vmware setup | 1 |
22,731 | 15,415,004,812 | IssuesEvent | 2021-03-05 01:34:46 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | Arm32 CI report xunit results | arch-arm32 area-Infrastructure-coreclr | Add logic to the arm_ci test script to print xunit results so jenkins can archive test failures correctly.
| 1.0 | Arm32 CI report xunit results - Add logic to the arm_ci test script to print xunit results so jenkins can archive test failures correctly.
| infrastructure | ci report xunit results add logic to the arm ci test script to print xunit results so jenkins can archive test failures correctly | 1 |
1,182 | 3,068,742,039 | IssuesEvent | 2015-08-18 17:01:43 | PATRIC3/patric3_website | https://api.github.com/repos/PATRIC3/patric3_website | closed | The special URL, http://tb.patricbrc.org, gives a 503 error | bug Infrastructure/Deployment | It used to redirect to https://www.patricbrc.org/portal/portal/patric/TB which is the special TB page we set up in PATRIC at NIAID's request. We need the http://tb.patricbrc.org functional again because that is the URL we distributed. Becky says this probably has something to do with the move to the cluster.
| 1.0 | The special URL, http://tb.patricbrc.org, gives a 503 error - It used to redirect to https://www.patricbrc.org/portal/portal/patric/TB which is the special TB page we set up in PATRIC at NIAID's request. We need the http://tb.patricbrc.org functional again because that is the URL we distributed. Becky says this probably has something to do with the move to the cluster.
| infrastructure | the special url gives a error it used to redirect to which is the special tb page we set up in patric at niaid s request we need the functional again because that is the url we distributed becky says this probably has something to do with the move to the cluster | 1 |
741,703 | 25,813,149,797 | IssuesEvent | 2022-12-12 01:25:46 | Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth-2 | https://api.github.com/repos/Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth-2 | opened | WC Ogre Duels Tradition | lore :books: priority medium :grey_exclamation: 2D graphics :paintbrush: cultural :mortar_board: | <!--
**DO NOT REMOVE PRE-EXISTING LINES**
------------------------------------------------------------------------------------------------------------
-->
### **Branched from #936 ; to be based on #880 after #936 is merged.**
Create a unique dueling WC tradition for Ogres to replace `tradition_fp1_trials_by_combat`.
| 1.0 | WC Ogre Duels Tradition - <!--
**DO NOT REMOVE PRE-EXISTING LINES**
------------------------------------------------------------------------------------------------------------
-->
### **Branched from #936 ; to be based on #880 after #936 is merged.**
Create a unique dueling WC tradition for Ogres to replace `tradition_fp1_trials_by_combat`.
| non_infrastructure | wc ogre duels tradition do not remove pre existing lines branched from to be based on after is merged create a unique dueling wc tradition for ogres to replace tradition trials by combat | 0 |
34,374 | 29,579,392,094 | IssuesEvent | 2023-06-07 03:39:18 | hackforla/food-oasis | https://api.github.com/repos/hackforla/food-oasis | closed | Bug: Container Build Failing | Release Note: Bug Role: Back-end Feature: Infrastructure | ### Overview
It appears that the docker container node:buster-slim is using a node version newer than LTS. When npm install attempts to install bcrypt, it fails, and attempts to build the bcrypt module from scratch, which also fails because the image does not have python installed to perform the build. Modify the docker container to an LTS version to resolve the problem.
### Action Items
- [ ] Modify the Dockerfile to use LTS node images.
- [ ] Deploy to test.
### Resources/Instructions
See [this article](https://snyk.io/blog/choosing-the-best-node-js-docker-image/) for an explanation of the node image choices.
| 1.0 | Bug: Container Build Failing - ### Overview
It appears that the docker container node:buster-slim is using a node version newer than LTS. When npm install attempts to install bcrypt, it fails, and attempts to build the bcrypt module from scratch, which also fails because the image does not have python installed to perform the build. Modify the docker container to an LTS version to resolve the problem.
### Action Items
- [ ] Modify the Dockerfile to use LTS node images.
- [ ] Deploy to test.
### Resources/Instructions
See [this article](https://snyk.io/blog/choosing-the-best-node-js-docker-image/) for an explanation of the node image choices.
| infrastructure | bug container build failing overview it appears that the docker container node buster slim is using a node version newer than lts when npm install attempts to install bcrypt it fails and attempts to build the bcrypt module from scratch which also fails because the image does not have python installed to perform the build modify the docker container to an lts version to resolve the problem action items modify the dockerfile to use lts node images deploy to test resources instructions see for an explanation of the node image choices | 1 |
69,886 | 22,714,991,660 | IssuesEvent | 2022-07-06 00:35:39 | zed-industries/feedback | https://api.github.com/repos/zed-industries/feedback | closed | Zed Crashes When Pasting into Settings | defect triage | **Before you begin**
Check the backlog of issues to reduce the chances of creating duplicates; if an issue already exists, place a `+1` (👍) on it.
**Describe the bug**
On a fresh start of Zed, I get a crash when I insert JSON into the settings.json file.
**To reproduce**
Steps to reproduce the behavior:
- Zed completely killed, settings.json is empty.
- Open Zed and press cmd-,
- Attempt to paste the following block:
```json
{
"autosave": {
"after_delay": {
"milliseconds": 1000
}
}
```
Via cmd-v or Edit->paste
- Zed should crash.
**Expected behavior**
Zed should not crash
**Screenshots**
https://user-images.githubusercontent.com/17581545/177408049-7ba0f659-ee5b-461a-be70-5d2a1b22594f.mov
**Environment:**
Zed 0.43.0 – /Applications/Zed.app
macOS 12.4
architecture arm64v
*If applicable, attach your `~/Library/Logs/Zed/Zed.log` file to this issue.*
| 1.0 | Zed Crashes When Pasting into Settings - **Before you begin**
Check the backlog of issues to reduce the chances of creating duplicates; if an issue already exists, place a `+1` (👍) on it.
**Describe the bug**
On a fresh start of Zed, I get a crash when I insert JSON into the settings.json file.
**To reproduce**
Steps to reproduce the behavior:
- Zed completely killed, settings.json is empty.
- Open Zed and press cmd-,
- Attempt to paste the following block:
```json
{
"autosave": {
"after_delay": {
"milliseconds": 1000
}
}
```
Via cmd-v or Edit->paste
- Zed should crash.
**Expected behavior**
Zed should not crash
**Screenshots**
https://user-images.githubusercontent.com/17581545/177408049-7ba0f659-ee5b-461a-be70-5d2a1b22594f.mov
**Environment:**
Zed 0.43.0 – /Applications/Zed.app
macOS 12.4
architecture arm64v
*If applicable, attach your `~/Library/Logs/Zed/Zed.log` file to this issue.*
| non_infrastructure | zed crashes when pasting into settings before you begin check the backlog of issues to reduce the chances of creating duplicates if an issue already exists place a 👍 on it describe the bug on a fresh start of zed i get a crash when i insert json into the settings json file to reproduce steps to reproduce the behavior zed completely killed settings json is empty open zed and press cmd attempt to paste the following block json autosave after delay milliseconds via cmd v or edit paste zed should crash expected behavior zed should not crash screenshots environment zed – applications zed app macos architecture if applicable attach your library logs zed zed log file to this issue | 0 |
7,040 | 6,720,142,210 | IssuesEvent | 2017-10-16 06:17:11 | wldcordeiro/cz-gitmoji-changelog | https://api.github.com/repos/wldcordeiro/cz-gitmoji-changelog | closed | TODO: Add CI and fix semantic-release | Infrastructure | Semantic-release is a cool tool but it needs CI setup (by default TravisCI). Get that up and running. | 1.0 | TODO: Add CI and fix semantic-release - Semantic-release is a cool tool but it needs CI setup (by default TravisCI). Get that up and running. | infrastructure | todo add ci and fix semantic release semantic release is a cool tool but it needs ci setup by default travisci get that up and running | 1 |
236,550 | 7,750,381,273 | IssuesEvent | 2018-05-30 14:13:58 | TwinePlatform/twine-visitor | https://api.github.com/repos/TwinePlatform/twine-visitor | opened | Fix all prop-type errors | chore priority-3 | Client has multiple prop-type errors, should be fixed to avoid polluting the console and make error monitoring less noisy. | 1.0 | Fix all prop-type errors - Client has multiple prop-type errors, should be fixed to avoid polluting the console and make error monitoring less noisy. | non_infrastructure | fix all prop type errors client has multiple prop type errors should be fixed to avoid polluting the console and make error monitoring less noisy | 0 |
313,019 | 23,451,287,465 | IssuesEvent | 2022-08-16 03:13:21 | SENG3150-Assignment-Group/3160-Front | https://api.github.com/repos/SENG3150-Assignment-Group/3160-Front | opened | Content required for the Admin, Agent, and Staff Pages | documentation enhancement requires validation | Before validating this issue please discuss here what content needs to be added to these pages in comments, preferably a sketch of what needs to be on each page before putting through any implementation (this also may need to be split up into multiple issues)
- Admin.jsx
- Agent.jsx
- Staff.jsx
- README.md document for each | 1.0 | Content required for the Admin, Agent, and Staff Pages - Before validating this issue please discuss here what content needs to be added to these pages in comments, preferably a sketch of what needs to be on each page before putting through any implementation (this also may need to be split up into multiple issues)
- Admin.jsx
- Agent.jsx
- Staff.jsx
- README.md document for each | non_infrastructure | content required for the admin agent and staff pages before validating this issue please discuss here what content needs to be added to these pages in comments preferably a sketch of what needs to be on each page before putting through any implementation this also may need to be split up into multiple issues admin jsx agent jsx staff jsx readme md document for each | 0 |
12,819 | 2,718,418,496 | IssuesEvent | 2015-04-12 09:00:52 | bridgedotnet/Bridge | https://api.github.com/repos/bridgedotnet/Bridge | closed | Unexpected string concatenation for System.Array.GetUpperBound(int) | defect | The C# code:
```csharp
using Bridge.Html5;
namespace ClassLibrary1
{
public class Class1
{
private static string[,] board = new string[3,3];
[Ready]
public static void main()
{
// FIXME: This returns NaN!?!?!?
Console.Log("Upper bounds: " + ClassLibrary1.Class1.board.GetUpperBound(0));
Console.Log("Upper bounds parenthesized: " + (ClassLibrary1.Class1.board.GetUpperBound(0)));
// Call ClassLibrary1.Class1.main() from console, it will show correct count.
}
}
}
```
results in
```javascript
Bridge.define('ClassLibrary1.Class1', {
statics: {
config: {
init: function () {
this.board = Bridge.Array.create(null, null, 3, 3);
Bridge.ready(this.main);
}
},
main: function () {
// FIXME: This returns NaN!?!?!?
console.log("Upper bounds: " + Bridge.Array.getLength(ClassLibrary1.Class1.board, 0) - 1);
console.log("Upper bounds parenthesized: " + (Bridge.Array.getLength(ClassLibrary1.Class1.board, 0) - 1));
// Call ClassLibrary1.Class1.main() from console, it will show correct count.
}
}
});
```
`System.Array.GetUpperBound(int)` resulting in a JavaScript expression makes an undesired side effect that can break in a number of situations. Instead, it would be a good idea to wrap JavaScript's `Bridge.Array.getLength()`, for example, with a `Bridge.Array.getUpperBound()` which internally does the subtraction.
Anything translated into an expression should be enclosed in parenthesis, for consistency. As it would not look great in every scenario, then maybe it would be a good idea to use the overloaded function.
This problem is somewhat similar to the issue #169, where adding additional parenthesis to the term fixes the problem.
**Notice:** `System.Array.GetLowerBound(int)` already has a bridge's overload `Bridge.Array.getLower()`, so this problem does not apply to this scenario. | 1.0 | Unexpected string concatenation for System.Array.GetUpperBound(int) - The C# code:
```csharp
using Bridge.Html5;
namespace ClassLibrary1
{
public class Class1
{
private static string[,] board = new string[3,3];
[Ready]
public static void main()
{
// FIXME: This returns NaN!?!?!?
Console.Log("Upper bounds: " + ClassLibrary1.Class1.board.GetUpperBound(0));
Console.Log("Upper bounds parenthesized: " + (ClassLibrary1.Class1.board.GetUpperBound(0)));
// Call ClassLibrary1.Class1.main() from console, it will show correct count.
}
}
}
```
results in
```javascript
Bridge.define('ClassLibrary1.Class1', {
statics: {
config: {
init: function () {
this.board = Bridge.Array.create(null, null, 3, 3);
Bridge.ready(this.main);
}
},
main: function () {
// FIXME: This returns NaN!?!?!?
console.log("Upper bounds: " + Bridge.Array.getLength(ClassLibrary1.Class1.board, 0) - 1);
console.log("Upper bounds parenthesized: " + (Bridge.Array.getLength(ClassLibrary1.Class1.board, 0) - 1));
// Call ClassLibrary1.Class1.main() from console, it will show correct count.
}
}
});
```
`System.Array.GetUpperBound(int)` resulting in a JavaScript expression makes an undesired side effect that can break in a number of situations. Instead, it would be a good idea to wrap JavaScript's `Bridge.Array.getLength()`, for example, with a `Bridge.Array.getUpperBound()` which internally does the subtraction.
Anything translated into an expression should be enclosed in parenthesis, for consistency. As it would not look great in every scenario, then maybe it would be a good idea to use the overloaded function.
This problem is somewhat similar to the issue #169, where adding additional parenthesis to the term fixes the problem.
**Notice:** `System.Array.GetLowerBound(int)` already has a bridge's overload `Bridge.Array.getLower()`, so this problem does not apply to this scenario. | non_infrastructure | unexpected string concatenation for system array getupperbound int the c code csharp using bridge namespace public class private static string board new string public static void main fixme this returns nan console log upper bounds board getupperbound console log upper bounds parenthesized board getupperbound call main from console it will show correct count results in javascript bridge define statics config init function this board bridge array create null null bridge ready this main main function fixme this returns nan console log upper bounds bridge array getlength board console log upper bounds parenthesized bridge array getlength board call main from console it will show correct count system array getupperbound int resulting in a javascript expression makes an undesired side effect that can break in a number of situations instead it would be a good idea to wrap javascript s bridge array getlength for example with a bridge array getupperbound which internally does the subtraction anything translated into an expression should be enclosed in parenthesis for consistency as it would not look great in every scenario then maybe it would be a good idea to use the overloaded function this problem is somewhat similar to the issue where adding additional parenthesis to the term fixes the problem notice system array getlowerbound int already has a bridge s overload bridge array getlower so this problem does not apply to this scenario | 0 |
704,453 | 24,197,716,411 | IssuesEvent | 2022-09-24 05:20:51 | mit-cml/appinventor-sources | https://api.github.com/repos/mit-cml/appinventor-sources | closed | Add menu item to remove all orphaned blocks | help wanted issue: noted for future Work status: forum feature request affects: master priority: low | **Describe the desired feature**
<!--
Describe the feature that you'd like to see implemented for App Inventor. More detail is useful as it allows us to better understand the complexity of the task.
-->
[From the forum](https://community.appinventor.mit.edu/t/no-way-to-mass-delete-orphaned-blocks/43110?u=ewpatton): During experimentation, one may accumulate many disconnected blocks that are not top level blocks. It may be desirable to clean up (i.e., delete) any non-declaration block (i.e., those blocks that aren't global definitions, procedure definitions, event handlers). It would be useful to add a new menu item to the Blockly workspace to delete any orphaned blocks. We should prompt to confirm the user intends to take this action, similar to deleting all of the blocks in the workspace.
Another potential design might be to provide this option behind a key combo. For example, on macOS you can often hold the option key to see additional menu items or alternative interpretations in the same spot. We might have "Delete all blocks (#### blocks)" and then when option is held show "Delete orphan blocks (#### blocks)". This would keep the context menu from becoming too cluttered.
**Give an example of how this feature would be used**
<!--
How would a teacher or student use this feature?
-->
The feature would be used to keep the code clean and workspace uncluttered.
**Why doesn't the current App Inventor system address this use case?**
<!--
Explain why the use case cannot be completed using the features of the current system.
-->
It does not provide this functionality. People can manually go around deleting unattached blocks (such as through the warning stepper buttons), but this may be cumbersome.
**Why is this feature beneficial to App Inventor's educational mission?**
<!--
Because MIT App Inventor is aimed at educational use, we prioritize development of features with an educational benefit. Help us understand how your feature request relates to our mission.
-->
The functionality can be used to help students maintain clean, understandable code during and after working on a project. | 1.0 | Add menu item to remove all orphaned blocks - **Describe the desired feature**
<!--
Describe the feature that you'd like to see implemented for App Inventor. More detail is useful as it allows us to better understand the complexity of the task.
-->
[From the forum](https://community.appinventor.mit.edu/t/no-way-to-mass-delete-orphaned-blocks/43110?u=ewpatton): During experimentation, one may accumulate many disconnected blocks that are not top level blocks. It may be desirable to clean up (i.e., delete) any non-declaration block (i.e., those blocks that aren't global definitions, procedure definitions, event handlers). It would be useful to add a new menu item to the Blockly workspace to delete any orphaned blocks. We should prompt to confirm the user intends to take this action, similar to deleting all of the blocks in the workspace.
Another potential design might be to provide this option behind a key combo. For example, on macOS you can often hold the option key to see additional menu items or alternative interpretations in the same spot. We might have "Delete all blocks (#### blocks)" and then when option is held show "Delete orphan blocks (#### blocks)". This would keep the context menu from becoming too cluttered.
**Give an example of how this feature would be used**
<!--
How would a teacher or student use this feature?
-->
The feature would be used to keep the code clean and workspace uncluttered.
**Why doesn't the current App Inventor system address this use case?**
<!--
Explain why the use case cannot be completed using the features of the current system.
-->
It does not provide this functionality. People can manually go around deleting unattached blocks (such as through the warning stepper buttons), but this may be cumbersome.
**Why is this feature beneficial to App Inventor's educational mission?**
<!--
Because MIT App Inventor is aimed at educational use, we prioritize development of features with an educational benefit. Help us understand how your feature request relates to our mission.
-->
The functionality can be used to help students maintain clean, understandable code during and after working on a project. | non_infrastructure | add menu item to remove all orphaned blocks describe the desired feature describe the feature that you d like to see implemented for app inventor more detail is useful as it allows us to better understand the complexity of the task during experimentation one may accumulate many disconnected blocks that are not top level blocks it may be desirable to clean up i e delete any non declaration block i e those blocks that aren t global definitions procedure definitions event handlers it would be useful to add a new menu item to the blockly workspace to delete any orphaned blocks we should prompt to confirm the user intends to take this action similar to deleting all of the blocks in the workspace another potential design might be to provide this option behind a key combo for example on macos you can often hold the option key to see additional menu items or alternative interpretations in the same spot we might have delete all blocks blocks and then when option is held show delete orphan blocks blocks this would keep the context menu from becoming too cluttered give an example of how this feature would be used how would a teacher or student use this feature the feature would be used to keep the code clean and workspace uncluttered why doesn t the current app inventor system address this use case explain why the use case cannot be completed using the features of the current system it does not provide this functionality people can manually go around deleting unattached blocks such as through the warning stepper buttons but this may be cumbersome why is this feature beneficial to app inventor s educational mission because mit app inventor is aimed at educational use we prioritize development of features with an educational benefit help us understand how your feature request relates to our mission the functionality can be used to help students maintain clean understandable code during and after working on a project | 0 |
401,387 | 11,789,638,880 | IssuesEvent | 2020-03-17 17:30:06 | airshipit/airshipctl | https://api.github.com/repos/airshipit/airshipctl | reopened | Remove kustomize related methods from Bundle interface | enhancement priority/critical ready for review | **Problem description (if applicable)**
Bundle interface right now offers following methods:
```
GetKustomizeResourceMap() resmap.ResMap
SetKustomizeResourceMap(resmap.ResMap) error
GetKustomizeBuildOptions() KustomizeBuildOptions
SetKustomizeBuildOptions(KustomizeBuildOptions) error
```
At this point these methods are only used internally by BundleFactory - bundle implementation.
Due to recent discussions about the future of kustomize, and possible shifts to kyaml or theoretical other backend for document aggregation engine, proposal is to remove these methods from Bundle interface, and instead move them to BundleFactory implementation of the interface.
**Proposed change**
Remove methods from Bundle interface and move them to BundleFactory implementation instead for internal use.
**Potential impacts**
At this stage there should be no impact, as these methods are not used by any other package | 1.0 | Remove kustomize related methods from Bundle interface - **Problem description (if applicable)**
Bundle interface right now offers following methods:
```
GetKustomizeResourceMap() resmap.ResMap
SetKustomizeResourceMap(resmap.ResMap) error
GetKustomizeBuildOptions() KustomizeBuildOptions
SetKustomizeBuildOptions(KustomizeBuildOptions) error
```
At this point these methods are only used internally by BundleFactory - bundle implementation.
Due to recent discussions about the future of kustomize, and possible shifts to kyaml or theoretical other backend for document aggregation engine, proposal is to remove these methods from Bundle interface, and instead move them to BundleFactory implementation of the interface.
**Proposed change**
Remove methods from Bundle interface and move them to BundleFactory implementation instead for internal use.
**Potential impacts**
At this stage there should be no impact, as these methods are not used by any other package | non_infrastructure | remove kustomize related methods from bundle interface problem description if applicable bundle interface right now offers following methods getkustomizeresourcemap resmap resmap setkustomizeresourcemap resmap resmap error getkustomizebuildoptions kustomizebuildoptions setkustomizebuildoptions kustomizebuildoptions error at this point these methods are only used internally by bundlefactory bundle implementation due to recent discussions about the future of kustomize and possible shifts to kyaml or theoretical other backend for document aggregation engine proposal is to remove these methods from bundle interface and instead move them to bundlefactory implementation of the interface proposed change remove methods from bundle interface and move them to bundlefactory implementation instead for internal use potential impacts at this stage there should be no impact as these methods are not used by any other package | 0 |
769,311 | 27,001,335,865 | IssuesEvent | 2023-02-10 08:04:57 | azerothcore/azerothcore-wotlk | https://api.github.com/repos/azerothcore/azerothcore-wotlk | closed | Perry Gatner does not seem to perform in the World's End Tavern. | Priority-Trivial 61-64 | ### Current Behaviour
No performance in the World's End Tavern.
### Expected Blizzlike Behaviour
Perry Gatner should perform stand-up comedy skits/routines in the tavern every so often, varying depending on the day. Shattrath Saul should also introduce Gatner beforehand.
### Source
[Wowpedia](https://wowpedia.fandom.com/wiki/Perry_Gatner?oldid=2517313)
### Steps to reproduce the problem
1. `.tele shattrath`
2. Go to the tavern.
3. Wait.
4. Observe no performance.
### Extra Notes
_No response_
### AC rev. hash/commit
e7e119283f776746f34f409480c567e87f078483
### Operating system
Windows 10 x64
### Custom changes or Modules
[mod-ptr-template](https://github.com/heyitsbench/mod-ptr-template) | 1.0 | Perry Gatner does not seem to perform in the World's End Tavern. - ### Current Behaviour
No performance in the World's End Tavern.
### Expected Blizzlike Behaviour
Perry Gatner should perform stand-up comedy skits/routines in the tavern every so often, varying depending on the day. Shattrath Saul should also introduce Gatner beforehand.
### Source
[Wowpedia](https://wowpedia.fandom.com/wiki/Perry_Gatner?oldid=2517313)
### Steps to reproduce the problem
1. `.tele shattrath`
2. Go to the tavern.
3. Wait.
4. Observe no performance.
### Extra Notes
_No response_
### AC rev. hash/commit
e7e119283f776746f34f409480c567e87f078483
### Operating system
Windows 10 x64
### Custom changes or Modules
[mod-ptr-template](https://github.com/heyitsbench/mod-ptr-template) | non_infrastructure | perry gatner does not seem to perform in the world s end tavern current behaviour no performance in the world s end tavern expected blizzlike behaviour perry gatner should perform stand up comedy skits routines in the tavern every so often varying depending on the day shattrath saul should also introduce gatner beforehand source steps to reproduce the problem tele shattrath go to the tavern wait observe no performance extra notes no response ac rev hash commit operating system windows custom changes or modules | 0 |
39,168 | 15,883,306,775 | IssuesEvent | 2021-04-09 17:12:19 | Seneca-CDOT/telescope | https://api.github.com/repos/Seneca-CDOT/telescope | closed | Migrate legacy accounts away from wiki | area: microservices area: tools | With the users microservice nearing completion we need a tool to download, parse, and send feed info from the wiki planet feed list to the users microservice via post requests.
The users microservice will store this data in firebase (or the emulator depending on the environment). | 1.0 | Migrate legacy accounts away from wiki - With the users microservice nearing completion we need a tool to download, parse, and send feed info from the wiki planet feed list to the users microservice via post requests.
The users microservice will store this data in firebase (or the emulator depending on the environment). | non_infrastructure | migrate legacy accounts away from wiki with the users microservice nearing completion we need a tool to download parse and send feed info from the wiki planet feed list to the users microservice via post requests the users microservice will store this data in firebase or the emulator depending on the environment | 0 |
343,622 | 10,333,874,140 | IssuesEvent | 2019-09-03 06:56:02 | wso2/product-is | https://api.github.com/repos/wso2/product-is | closed | SQLSyntaxErrorException when updating the Callback Url of the OAuth/OpenID Connect Service Provider | Priority/High Severity/Critical Type/Bug WUM | When updating the Callback Url of the OAuth/OpenID Connect Service Provider, getting MySQLSyntaxErrorException.
The SQL query engaged in this operation is malformed (a comma is missing near ID_TOKEN_EXPIRE_TIME=?).
`public static final String UPDATE_CONSUMER_APP_WITH_OWNER_UPDATE = "UPDATE IDN_OAUTH_CONSUMER_APPS SET APP_NAME=?," +
" CALLBACK_URL=?, GRANT_TYPES=?, USER_ACCESS_TOKEN_EXPIRE_TIME=?, " +
"APP_ACCESS_TOKEN_EXPIRE_TIME=?, REFRESH_TOKEN_EXPIRE_TIME=?, ID_TOKEN_EXPIRE_TIME=? USERNAME=?, USER_DOMAIN=?" +
"WHERE CONSUMER_KEY=?";`
In the same SQL query, space is missing before the WHERE clause. It seems like not affecting the MySQL DB but may affect other DB types.
The above syntax error persists in **IS-5.7.0** version onwards. Please refer [1].
[1] https://github.com/wso2-extensions/identity-inbound-auth-oauth/blob/master/components/org.wso2.carbon.identity.oauth/src/main/java/org/wso2/carbon/identity/oauth/dao/SQLQueries.java#L119 | 1.0 | SQLSyntaxErrorException when updating the Callback Url of the OAuth/OpenID Connect Service Provider - When updating the Callback Url of the OAuth/OpenID Connect Service Provider, getting MySQLSyntaxErrorException.
The SQL query engaged in this operation is malformed (a comma is missing near ID_TOKEN_EXPIRE_TIME=?).
`public static final String UPDATE_CONSUMER_APP_WITH_OWNER_UPDATE = "UPDATE IDN_OAUTH_CONSUMER_APPS SET APP_NAME=?," +
" CALLBACK_URL=?, GRANT_TYPES=?, USER_ACCESS_TOKEN_EXPIRE_TIME=?, " +
"APP_ACCESS_TOKEN_EXPIRE_TIME=?, REFRESH_TOKEN_EXPIRE_TIME=?, ID_TOKEN_EXPIRE_TIME=? USERNAME=?, USER_DOMAIN=?" +
"WHERE CONSUMER_KEY=?";`
In the same SQL query, space is missing before the WHERE clause. It seems like not affecting the MySQL DB but may affect other DB types.
The above syntax error persists in **IS-5.7.0** version onwards. Please refer [1].
[1] https://github.com/wso2-extensions/identity-inbound-auth-oauth/blob/master/components/org.wso2.carbon.identity.oauth/src/main/java/org/wso2/carbon/identity/oauth/dao/SQLQueries.java#L119 | non_infrastructure | sqlsyntaxerrorexception when updating the callback url of the oauth openid connect service provider when updating the callback url of the oauth openid connect service provider getting mysqlsyntaxerrorexception the sql query engaged in this operation is malformed a comma is missing near id token expire time public static final string update consumer app with owner update update idn oauth consumer apps set app name callback url grant types user access token expire time app access token expire time refresh token expire time id token expire time username user domain where consumer key in the same sql query space is missing before the where clause it seems like not affecting the mysql db but may affect other db types the above syntax error persists in is version onwards please refer | 0 |
54,145 | 23,171,586,691 | IssuesEvent | 2022-07-30 20:04:33 | imneverland/upptime | https://api.github.com/repos/imneverland/upptime | opened | 🛑 NeverLand Hosting Services is down | status never-land-hosting-services | In [`3b86fad`](https://github.com/imneverland/upptime/commit/3b86fad60ff06d2af1da519a3ca5f6d6450bb9b5
), NeverLand Hosting Services ($HOSTING_IP) was **down**:
- HTTP code: 0
- Response time: 0 ms
| 1.0 | 🛑 NeverLand Hosting Services is down - In [`3b86fad`](https://github.com/imneverland/upptime/commit/3b86fad60ff06d2af1da519a3ca5f6d6450bb9b5
), NeverLand Hosting Services ($HOSTING_IP) was **down**:
- HTTP code: 0
- Response time: 0 ms
| non_infrastructure | 🛑 neverland hosting services is down in neverland hosting services hosting ip was down http code response time ms | 0 |
100,490 | 16,489,891,892 | IssuesEvent | 2021-05-25 01:06:36 | billmcchesney1/react-bootstrap | https://api.github.com/repos/billmcchesney1/react-bootstrap | opened | CVE-2021-23362 (Medium) detected in hosted-git-info-2.8.8.tgz, hosted-git-info-3.0.7.tgz | security vulnerability | ## CVE-2021-23362 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>hosted-git-info-2.8.8.tgz</b>, <b>hosted-git-info-3.0.7.tgz</b></p></summary>
<p>
<details><summary><b>hosted-git-info-2.8.8.tgz</b></p></summary>
<p>Provides metadata and conversions from repository urls for Github, Bitbucket and Gitlab</p>
<p>Library home page: <a href="https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz">https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz</a></p>
<p>Path to dependency file: react-bootstrap/package.json</p>
<p>Path to vulnerable library: react-bootstrap/node_modules/conventional-changelog-core/node_modules/read-pkg/node_modules/hosted-git-info/package.json,react-bootstrap/www/node_modules/hosted-git-info/package.json,react-bootstrap/node_modules/hosted-git-info/package.json</p>
<p>
Dependency Hierarchy:
- node-sass-4.14.1.tgz (Root Library)
- meow-3.7.0.tgz
- normalize-package-data-2.5.0.tgz
- :x: **hosted-git-info-2.8.8.tgz** (Vulnerable Library)
</details>
<details><summary><b>hosted-git-info-3.0.7.tgz</b></p></summary>
<p>Provides metadata and conversions from repository urls for Github, Bitbucket and Gitlab</p>
<p>Library home page: <a href="https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-3.0.7.tgz">https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-3.0.7.tgz</a></p>
<p>Path to dependency file: react-bootstrap/www/package.json</p>
<p>Path to vulnerable library: react-bootstrap/www/node_modules/gatsby/node_modules/hosted-git-info/package.json,react-bootstrap/node_modules/meow/node_modules/hosted-git-info/package.json,react-bootstrap/node_modules/conventional-changelog-core/node_modules/hosted-git-info/package.json</p>
<p>
Dependency Hierarchy:
- conventional-changelog-cli-2.1.1.tgz (Root Library)
- meow-8.1.0.tgz
- normalize-package-data-3.0.0.tgz
- :x: **hosted-git-info-3.0.7.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity.
<p>Publish Date: 2021-03-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362>CVE-2021-23362</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-43f8-2h32-f4cj">https://github.com/advisories/GHSA-43f8-2h32-f4cj</a></p>
<p>Release Date: 2021-03-23</p>
<p>Fix Resolution: hosted-git-info - 2.8.9,3.0.8</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"hosted-git-info","packageVersion":"2.8.8","packageFilePaths":["/package.json","/www/package.json"],"isTransitiveDependency":true,"dependencyTree":"node-sass:4.14.1;meow:3.7.0;normalize-package-data:2.5.0;hosted-git-info:2.8.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"hosted-git-info - 2.8.9,3.0.8"},{"packageType":"javascript/Node.js","packageName":"hosted-git-info","packageVersion":"3.0.7","packageFilePaths":["/www/package.json","/package.json"],"isTransitiveDependency":true,"dependencyTree":"conventional-changelog-cli:2.1.1;meow:8.1.0;normalize-package-data:3.0.0;hosted-git-info:3.0.7","isMinimumFixVersionAvailable":true,"minimumFixVersion":"hosted-git-info - 2.8.9,3.0.8"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-23362","vulnerabilityDetails":"The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2021-23362 (Medium) detected in hosted-git-info-2.8.8.tgz, hosted-git-info-3.0.7.tgz - ## CVE-2021-23362 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>hosted-git-info-2.8.8.tgz</b>, <b>hosted-git-info-3.0.7.tgz</b></p></summary>
<p>
<details><summary><b>hosted-git-info-2.8.8.tgz</b></p></summary>
<p>Provides metadata and conversions from repository urls for Github, Bitbucket and Gitlab</p>
<p>Library home page: <a href="https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz">https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-2.8.8.tgz</a></p>
<p>Path to dependency file: react-bootstrap/package.json</p>
<p>Path to vulnerable library: react-bootstrap/node_modules/conventional-changelog-core/node_modules/read-pkg/node_modules/hosted-git-info/package.json,react-bootstrap/www/node_modules/hosted-git-info/package.json,react-bootstrap/node_modules/hosted-git-info/package.json</p>
<p>
Dependency Hierarchy:
- node-sass-4.14.1.tgz (Root Library)
- meow-3.7.0.tgz
- normalize-package-data-2.5.0.tgz
- :x: **hosted-git-info-2.8.8.tgz** (Vulnerable Library)
</details>
<details><summary><b>hosted-git-info-3.0.7.tgz</b></p></summary>
<p>Provides metadata and conversions from repository urls for Github, Bitbucket and Gitlab</p>
<p>Library home page: <a href="https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-3.0.7.tgz">https://registry.npmjs.org/hosted-git-info/-/hosted-git-info-3.0.7.tgz</a></p>
<p>Path to dependency file: react-bootstrap/www/package.json</p>
<p>Path to vulnerable library: react-bootstrap/www/node_modules/gatsby/node_modules/hosted-git-info/package.json,react-bootstrap/node_modules/meow/node_modules/hosted-git-info/package.json,react-bootstrap/node_modules/conventional-changelog-core/node_modules/hosted-git-info/package.json</p>
<p>
Dependency Hierarchy:
- conventional-changelog-cli-2.1.1.tgz (Root Library)
- meow-8.1.0.tgz
- normalize-package-data-3.0.0.tgz
- :x: **hosted-git-info-3.0.7.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity.
<p>Publish Date: 2021-03-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362>CVE-2021-23362</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-43f8-2h32-f4cj">https://github.com/advisories/GHSA-43f8-2h32-f4cj</a></p>
<p>Release Date: 2021-03-23</p>
<p>Fix Resolution: hosted-git-info - 2.8.9,3.0.8</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"hosted-git-info","packageVersion":"2.8.8","packageFilePaths":["/package.json","/www/package.json"],"isTransitiveDependency":true,"dependencyTree":"node-sass:4.14.1;meow:3.7.0;normalize-package-data:2.5.0;hosted-git-info:2.8.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"hosted-git-info - 2.8.9,3.0.8"},{"packageType":"javascript/Node.js","packageName":"hosted-git-info","packageVersion":"3.0.7","packageFilePaths":["/www/package.json","/package.json"],"isTransitiveDependency":true,"dependencyTree":"conventional-changelog-cli:2.1.1;meow:8.1.0;normalize-package-data:3.0.0;hosted-git-info:3.0.7","isMinimumFixVersionAvailable":true,"minimumFixVersion":"hosted-git-info - 2.8.9,3.0.8"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2021-23362","vulnerabilityDetails":"The package hosted-git-info before 3.0.8 are vulnerable to Regular Expression Denial of Service (ReDoS) via regular expression shortcutMatch in the fromUrl function in index.js. The affected regular expression exhibits polynomial worst-case time complexity.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_infrastructure | cve medium detected in hosted git info tgz hosted git info tgz cve medium severity vulnerability vulnerable libraries hosted git info tgz hosted git info tgz hosted git info tgz provides metadata and conversions from repository urls for github bitbucket and gitlab library home page a href path to dependency file react bootstrap package json path to vulnerable library react bootstrap node modules conventional changelog core node modules read pkg node modules hosted git info package json react bootstrap www node modules hosted git info package json react bootstrap node modules hosted git info package json dependency hierarchy node sass tgz root library meow tgz normalize package data tgz x hosted git info tgz vulnerable library hosted git info tgz provides metadata and conversions from repository urls for github bitbucket and gitlab library home page a href path to dependency file react bootstrap www package json path to vulnerable library react bootstrap www node modules gatsby node modules hosted git info package json react bootstrap node modules meow node modules hosted git info package json react bootstrap node modules conventional changelog core node modules hosted git info package json dependency hierarchy conventional changelog cli tgz root library meow tgz normalize package data tgz x hosted git info tgz vulnerable library found in base branch master vulnerability details the package hosted git info before are vulnerable to regular expression denial of service redos via regular expression shortcutmatch in the fromurl function in index js the affected regular expression exhibits polynomial worst case time complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution hosted git info isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree node sass meow normalize package data hosted git info isminimumfixversionavailable true minimumfixversion hosted git info packagetype javascript node js packagename hosted git info packageversion packagefilepaths istransitivedependency true dependencytree conventional changelog cli meow normalize package data hosted git info isminimumfixversionavailable true minimumfixversion hosted git info basebranches vulnerabilityidentifier cve vulnerabilitydetails the package hosted git info before are vulnerable to regular expression denial of service redos via regular expression shortcutmatch in the fromurl function in index js the affected regular expression exhibits polynomial worst case time complexity vulnerabilityurl | 0 |
552,919 | 16,330,755,965 | IssuesEvent | 2021-05-12 08:56:13 | hotg-ai/rune | https://api.github.com/repos/hotg-ai/rune | closed | For image capability, need to extract from manifest what imagesize is, if its RGB BGR YUV or Grayscale | category - enhancement priority - urgent | Need these parameters to reformat in app.
Needs to be send from
manifest -> SDK -> App
Links to https://github.com/hotg-ai/rune_vm/issues/3 | 1.0 | For image capability, need to extract from manifest what imagesize is, if its RGB BGR YUV or Grayscale - Need these parameters to reformat in app.
Needs to be send from
manifest -> SDK -> App
Links to https://github.com/hotg-ai/rune_vm/issues/3 | non_infrastructure | for image capability need to extract from manifest what imagesize is if its rgb bgr yuv or grayscale need these parameters to reformat in app needs to be send from manifest sdk app links to | 0 |
11,445 | 9,200,530,928 | IssuesEvent | 2019-03-07 17:16:26 | elastic/beats | https://api.github.com/repos/elastic/beats | closed | Jolokia module failed to parse response for some metrics with GET method | :infrastructure Metricbeat bug module | Jolokia module fails to parse some responses fetched using the GET method with the error:
```
json: cannot unmarshal number into Go struct field Entry.Value of type map[string]interface {}
failed to unmarshal jolokia JSON response '{"request":{"mbean":"java.lang:type=Runtime","attribute":"Uptime","type":"read"},"value":88622,"timestamp":1551739190,"status":200}'
```
For confirmed bugs, please report:
- Version: >= 6.6.0 (only seems to affect GET method)
- Steps to Reproduce:
Use this config with the jolokia module:
```
http_method: 'GET'
jmx.mappings:
- mbean: 'java.lang:type=Runtime'
attributes:
- attr: Uptime
field: uptime
```
Reproduced also adding this case to the Jolokia integration tests:
```
{
"mbean": "java.lang:type=Runtime",
"attributes": []map[string]string{
{
"attr": "Uptime",
"field": "uptime",
},
},
},
``` | 1.0 | Jolokia module failed to parse response for some metrics with GET method - Jolokia module fails to parse some responses fetched using the GET method with the error:
```
json: cannot unmarshal number into Go struct field Entry.Value of type map[string]interface {}
failed to unmarshal jolokia JSON response '{"request":{"mbean":"java.lang:type=Runtime","attribute":"Uptime","type":"read"},"value":88622,"timestamp":1551739190,"status":200}'
```
For confirmed bugs, please report:
- Version: >= 6.6.0 (only seems to affect GET method)
- Steps to Reproduce:
Use this config with the jolokia module:
```
http_method: 'GET'
jmx.mappings:
- mbean: 'java.lang:type=Runtime'
attributes:
- attr: Uptime
field: uptime
```
Reproduced also adding this case to the Jolokia integration tests:
```
{
"mbean": "java.lang:type=Runtime",
"attributes": []map[string]string{
{
"attr": "Uptime",
"field": "uptime",
},
},
},
``` | infrastructure | jolokia module failed to parse response for some metrics with get method jolokia module fails to parse some responses fetched using the get method with the error json cannot unmarshal number into go struct field entry value of type map interface failed to unmarshal jolokia json response request mbean java lang type runtime attribute uptime type read value timestamp status for confirmed bugs please report version only seems to affect get method steps to reproduce use this config with the jolokia module http method get jmx mappings mbean java lang type runtime attributes attr uptime field uptime reproduced also adding this case to the jolokia integration tests mbean java lang type runtime attributes map string attr uptime field uptime | 1 |
735,086 | 25,378,729,975 | IssuesEvent | 2022-11-21 15:54:59 | netdata/netdata-cloud | https://api.github.com/repos/netdata/netdata-cloud | opened | [Bug]: The "net" section in the cgroups container is picking the wrong / random interface name | bug priority/high cloud-frontend needs triage | ### Bug description
The net section under cgroups (container) adds a random interface to the sub-section name (probably from one of the containers).
### Expected behavior
The sub-section should only be named generically as "net".
### Steps to reproduce
1. Create multiple containers on the system with different virtual interfaces
2. Open the Cloud dashboard and look at the cgroups container section.
### Screenshots
<img width="245" alt="image" src="https://user-images.githubusercontent.com/96257330/203098975-06214466-c17c-414a-8d9d-9b3136d6bf31.png">
For example, the net eth0 section in the image should be named as "net" only - especially for cases where there are multiple virtual interfaces belonging to different containers.
### Error Logs
_No response_
### Desktop
This is agent independent.
### Additional context
_No response_ | 1.0 | [Bug]: The "net" section in the cgroups container is picking the wrong / random interface name - ### Bug description
The net section under cgroups (container) adds a random interface to the sub-section name (probably from one of the containers).
### Expected behavior
The sub-section should only be named generically as "net".
### Steps to reproduce
1. Create multiple containers on the system with different virtual interfaces
2. Open the Cloud dashboard and look at the cgroups container section.
### Screenshots
<img width="245" alt="image" src="https://user-images.githubusercontent.com/96257330/203098975-06214466-c17c-414a-8d9d-9b3136d6bf31.png">
For example, the net eth0 section in the image should be named as "net" only - especially for cases where there are multiple virtual interfaces belonging to different containers.
### Error Logs
_No response_
### Desktop
This is agent independent.
### Additional context
_No response_ | non_infrastructure | the net section in the cgroups container is picking the wrong random interface name bug description the net section under cgroups container adds a random interface to the sub section name probably from one of the containers expected behavior the sub section should only be named generically as net steps to reproduce create multiple containers on the system with different virtual interfaces open the cloud dashboard and look at the cgroups container section screenshots img width alt image src for example the net section in the image should be named as net only especially for cases where there are multiple virtual interfaces belonging to different containers error logs no response desktop this is agent independent additional context no response | 0 |
5,575 | 20,144,032,037 | IssuesEvent | 2022-02-09 04:25:02 | theglus/Home-Assistant-Config | https://api.github.com/repos/theglus/Home-Assistant-Config | opened | Refine Dyson AQI automation | automation climate | # Requirements
- [ ] Create automation to turn-off kitchen Dyson when air quality is below defined threshold.
- [ ] Split out `scene.dyson_aqi` from `scene.yaml`.
- [ ] Split out `automation.climate_aqi` from `automations.yaml`. | 1.0 | Refine Dyson AQI automation - # Requirements
- [ ] Create automation to turn-off kitchen Dyson when air quality is below defined threshold.
- [ ] Split out `scene.dyson_aqi` from `scene.yaml`.
- [ ] Split out `automation.climate_aqi` from `automations.yaml`. | non_infrastructure | refine dyson aqi automation requirements create automation to turn off kitchen dyson when air quality is below defined threshold split out scene dyson aqi from scene yaml split out automation climate aqi from automations yaml | 0 |
651,417 | 21,477,734,544 | IssuesEvent | 2022-04-26 14:56:58 | ontox-hu/aspis4j | https://api.github.com/repos/ontox-hu/aspis4j | closed | add ssl to toxtrack.ddns.net | high-priority | 1. [x] serve neo4j through nginx on toxtrack.ddns.net
3. [x] add ssl certificate
4. [x] load production database
toxtrack.ddns.net will serve as the production version of aspis4j for now. | 1.0 | add ssl to toxtrack.ddns.net - 1. [x] serve neo4j through nginx on toxtrack.ddns.net
3. [x] add ssl certificate
4. [x] load production database
toxtrack.ddns.net will serve as the production version of aspis4j for now. | non_infrastructure | add ssl to toxtrack ddns net serve through nginx on toxtrack ddns net add ssl certificate load production database toxtrack ddns net will serve as the production version of for now | 0 |
14,702 | 11,054,044,166 | IssuesEvent | 2019-12-10 12:42:25 | code4romania/expert-consultation-api | https://api.github.com/repos/code4romania/expert-consultation-api | closed | [Infrastructure] Automatic deployment on AWS | docker enhancement good first issue infrastructure | As part of easy deployment for the application, we need to implement automatic deployment of the docker image containing the project and it's dependencies on a AWS machine. As an automatic deployment provider we will be using Travis-CI. | 1.0 | [Infrastructure] Automatic deployment on AWS - As part of easy deployment for the application, we need to implement automatic deployment of the docker image containing the project and it's dependencies on a AWS machine. As an automatic deployment provider we will be using Travis-CI. | infrastructure | automatic deployment on aws as part of easy deployment for the application we need to implement automatic deployment of the docker image containing the project and it s dependencies on a aws machine as an automatic deployment provider we will be using travis ci | 1 |
181,598 | 14,059,376,526 | IssuesEvent | 2020-11-03 02:55:27 | pingcap/tidb | https://api.github.com/repos/pingcap/tidb | closed | placement_sql_test.go:testDBSuite1.TestPlacementPolicyCache failed | component/test status/help-wanted | placement_sql_test.go:testDBSuite1.TestPlacementPolicyCache
```
[2020-11-02T10:58:53.976Z] ----------------------------------------------------------------------
[2020-11-02T10:58:53.976Z] FAIL: placement_sql_test.go:279: testDBSuite1.TestPlacementPolicyCache
[2020-11-02T10:58:53.976Z]
[2020-11-02T10:58:53.976Z] placement_sql_test.go:316:
[2020-11-02T10:58:53.976Z] tk.MustQuery("select * from information_schema.placement_policy").Check(testkit.Rows(rows...))
[2020-11-02T10:58:53.976Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/util/testkit/testkit.go:63:
[2020-11-02T10:58:53.976Z] res.c.Assert(resBuff.String(), check.Equals, needBuff.String(), res.comment)
[2020-11-02T10:58:53.976Z] ... obtained string = "" +
[2020-11-02T10:58:53.976Z] ... "[TiDB_DDL_487 0 default test t1 p1 <nil> 0 ]\n" +
[2020-11-02T10:58:53.976Z] ... "[TiDB_DDL_486 0 default test t1 p0 <nil> 0 ]\n"
[2020-11-02T10:58:53.976Z] ... expected string = "" +
[2020-11-02T10:58:53.976Z] ... "[TiDB_DDL_486 0 default test t1 p0 <nil> 0 ]\n" +
[2020-11-02T10:58:53.976Z] ... "[TiDB_DDL_487 0 default test t1 p1 <nil> 0 ]\n"
[2020-11-02T10:58:53.976Z] ... sql:select * from information_schema.placement_policy, args:[]
```
Latest failed builds:
https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_unit_test/56573/display/redirect
https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_check_2/58341/display/redirect
https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_check_2/58349/display/redirect
https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_check_2/58325/display/redirect
https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_check_2/58364/display/redirect
https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_check_2/58365/display/redirect
| 1.0 | placement_sql_test.go:testDBSuite1.TestPlacementPolicyCache failed - placement_sql_test.go:testDBSuite1.TestPlacementPolicyCache
```
[2020-11-02T10:58:53.976Z] ----------------------------------------------------------------------
[2020-11-02T10:58:53.976Z] FAIL: placement_sql_test.go:279: testDBSuite1.TestPlacementPolicyCache
[2020-11-02T10:58:53.976Z]
[2020-11-02T10:58:53.976Z] placement_sql_test.go:316:
[2020-11-02T10:58:53.976Z] tk.MustQuery("select * from information_schema.placement_policy").Check(testkit.Rows(rows...))
[2020-11-02T10:58:53.976Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/util/testkit/testkit.go:63:
[2020-11-02T10:58:53.976Z] res.c.Assert(resBuff.String(), check.Equals, needBuff.String(), res.comment)
[2020-11-02T10:58:53.976Z] ... obtained string = "" +
[2020-11-02T10:58:53.976Z] ... "[TiDB_DDL_487 0 default test t1 p1 <nil> 0 ]\n" +
[2020-11-02T10:58:53.976Z] ... "[TiDB_DDL_486 0 default test t1 p0 <nil> 0 ]\n"
[2020-11-02T10:58:53.976Z] ... expected string = "" +
[2020-11-02T10:58:53.976Z] ... "[TiDB_DDL_486 0 default test t1 p0 <nil> 0 ]\n" +
[2020-11-02T10:58:53.976Z] ... "[TiDB_DDL_487 0 default test t1 p1 <nil> 0 ]\n"
[2020-11-02T10:58:53.976Z] ... sql:select * from information_schema.placement_policy, args:[]
```
Latest failed builds:
https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_unit_test/56573/display/redirect
https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_check_2/58341/display/redirect
https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_check_2/58349/display/redirect
https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_check_2/58325/display/redirect
https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_check_2/58364/display/redirect
https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_check_2/58365/display/redirect
| non_infrastructure | placement sql test go testplacementpolicycache failed placement sql test go testplacementpolicycache fail placement sql test go testplacementpolicycache placement sql test go tk mustquery select from information schema placement policy check testkit rows rows home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb util testkit testkit go res c assert resbuff string check equals needbuff string res comment obtained string n n expected string n n sql select from information schema placement policy args latest failed builds | 0 |
1,493 | 4,018,772,857 | IssuesEvent | 2016-05-16 12:20:14 | AdguardTeam/AdguardForWindows | https://api.github.com/repos/AdguardTeam/AdguardForWindows | opened | [Win10 Insider preview 14332] Microsoft Edge crashes when right-clicking on highlighted text | Bug Compatibility | "Upon right clicking any highlighted text within Edge, the browser crashes almost immediately after hanging for a second"
Please refer to LGQ-473-77181 | True | [Win10 Insider preview 14332] Microsoft Edge crashes when right-clicking on highlighted text - "Upon right clicking any highlighted text within Edge, the browser crashes almost immediately after hanging for a second"
Please refer to LGQ-473-77181 | non_infrastructure | microsoft edge crashes when right clicking on highlighted text upon right clicking any highlighted text within edge the browser crashes almost immediately after hanging for a second please refer to lgq | 0 |
14,298 | 10,739,854,281 | IssuesEvent | 2019-10-29 17:05:22 | HumanCellAtlas/secondary-analysis | https://api.github.com/repos/HumanCellAtlas/secondary-analysis | closed | Monitor Falcon | Q1-2019 devops infrastructure | AC
1. Have a way to monitor the threads of Falcon in Grafana
┆Issue is synchronized with this [Jira Story](https://broadinstitute.atlassian.net/browse/GH-398)
| 1.0 | Monitor Falcon - AC
1. Have a way to monitor the threads of Falcon in Grafana
┆Issue is synchronized with this [Jira Story](https://broadinstitute.atlassian.net/browse/GH-398)
| infrastructure | monitor falcon ac have a way to monitor the threads of falcon in grafana ┆issue is synchronized with this | 1 |
127,997 | 18,024,845,621 | IssuesEvent | 2021-09-17 02:10:48 | faizulho/vuepress-deploy | https://api.github.com/repos/faizulho/vuepress-deploy | opened | CVE-2021-3801 (Medium) detected in prismjs-1.14.0.tgz | security vulnerability | ## CVE-2021-3801 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>prismjs-1.14.0.tgz</b></p></summary>
<p>Lightweight, robust, elegant syntax highlighting. A spin-off project from Dabblet.</p>
<p>Library home page: <a href="https://registry.npmjs.org/prismjs/-/prismjs-1.14.0.tgz">https://registry.npmjs.org/prismjs/-/prismjs-1.14.0.tgz</a></p>
<p>Path to dependency file: vuepress-deploy/package.json</p>
<p>Path to vulnerable library: vuepress-deploy/node_modules/prismjs/package.json</p>
<p>
Dependency Hierarchy:
- vuepress-0.6.1.tgz (Root Library)
- :x: **prismjs-1.14.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
prism is vulnerable to Inefficient Regular Expression Complexity
<p>Publish Date: 2021-09-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3801>CVE-2021-3801</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: N/A
- Attack Complexity: N/A
- Privileges Required: N/A
- User Interaction: N/A
- Scope: N/A
- Impact Metrics:
- Confidentiality Impact: N/A
- Integrity Impact: N/A
- Availability Impact: N/A
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-3801 (Medium) detected in prismjs-1.14.0.tgz - ## CVE-2021-3801 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>prismjs-1.14.0.tgz</b></p></summary>
<p>Lightweight, robust, elegant syntax highlighting. A spin-off project from Dabblet.</p>
<p>Library home page: <a href="https://registry.npmjs.org/prismjs/-/prismjs-1.14.0.tgz">https://registry.npmjs.org/prismjs/-/prismjs-1.14.0.tgz</a></p>
<p>Path to dependency file: vuepress-deploy/package.json</p>
<p>Path to vulnerable library: vuepress-deploy/node_modules/prismjs/package.json</p>
<p>
Dependency Hierarchy:
- vuepress-0.6.1.tgz (Root Library)
- :x: **prismjs-1.14.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
prism is vulnerable to Inefficient Regular Expression Complexity
<p>Publish Date: 2021-09-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3801>CVE-2021-3801</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: N/A
- Attack Complexity: N/A
- Privileges Required: N/A
- User Interaction: N/A
- Scope: N/A
- Impact Metrics:
- Confidentiality Impact: N/A
- Integrity Impact: N/A
- Availability Impact: N/A
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | cve medium detected in prismjs tgz cve medium severity vulnerability vulnerable library prismjs tgz lightweight robust elegant syntax highlighting a spin off project from dabblet library home page a href path to dependency file vuepress deploy package json path to vulnerable library vuepress deploy node modules prismjs package json dependency hierarchy vuepress tgz root library x prismjs tgz vulnerable library found in base branch master vulnerability details prism is vulnerable to inefficient regular expression complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector n a attack complexity n a privileges required n a user interaction n a scope n a impact metrics confidentiality impact n a integrity impact n a availability impact n a for more information on scores click a href step up your open source security game with whitesource | 0 |
28,139 | 23,041,912,192 | IssuesEvent | 2022-07-23 09:16:30 | ansible-collections/community.general | https://api.github.com/repos/ansible-collections/community.general | closed | Fix wrong url path in plugins/modules/web_infrastructure/jenkins_plugin.py | bug module plugins web_infrastructure | ### Summary
When we use option from community.general.jenkins_plugin module to verify installed plugins. We received error like below:
`failed: [localhost] (item={'key': 'http_request', 'value': '1.15'}) => {"ansible_loop_var": "item", "attempts": 3, "changed": false, "details": "HTTP Error 404: Not Found", "item": {"key": "http_request", "value": "1.15"}, "msg": "Plugin not found. http://localhost:8080/pluginManager/plugin/http_request/pin"}`
There is wrong url path "pluginManager/plugin" **Which need to be modified to** "pluginManager/plugins"
Follow link below:
https://github.com/ansible-collections/community.general/blob/788cfb624aa7e694b533b5d2722aab0ff4522ce6/plugins/modules/web_infrastructure/jenkins_plugin.py#L774
### Issue Type
Bug Report
### Component Name
plugins/modules/web_infrastructure/jenkins_plugin.py
### Ansible Version
```console (paste below)
sh-4.2$ ansible --version
[DEPRECATION WARNING]: Ansible will require Python 3.8 or newer on the controller starting with Ansible 2.12. Current version: 3.7.10 (default, Jun 3 2021, 00:02:01) [GCC 7.3.1
20180712 (Red Hat 7.3.1-13)]. This feature will be removed from ansible-core in version 2.12. Deprecation warnings can be disabled by setting deprecation_warnings=False in
ansible.cfg.
ansible [core 2.11.12]
config file = /etc/ansible/ansible.cfg
configured module search path = ['/home/ssm-user/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
ansible python module location = /usr/local/lib/python3.7/site-packages/ansible
ansible collection location = /home/ssm-user/.ansible/collections:/usr/share/ansible/collections
executable location = /usr/local/bin/ansible
python version = 3.7.10 (default, Jun 3 2021, 00:02:01) [GCC 7.3.1 20180712 (Red Hat 7.3.1-13)]
jinja version = 3.1.2
libyaml = True
```
### Community.general Version
```console (paste below)
$ ansible-galaxy collection list community.general
# /usr/local/lib/python3.7/site-packages/ansible_collections
Collection Version
----------------- -------
community.general 3.8.3
```
### Configuration
```console (paste below)
$ ansible-config dump --only-changed
```
### OS / Environment
_No response_
### Steps to Reproduce
<!--- Paste example playbooks or commands between quotes below -->
```yaml (paste below)
#Simple http request on Jenkins controller
sh-4.2$ curl http://localhost:8080/pluginManager/plugin
<html>
<head>
<meta http-equiv="Content-Type" content="text/html;charset=utf-8"/>
<title>Error 404 Not Found</title>
</head>
<body><h2>HTTP ERROR 404 Not Found</h2>
<table>
<tr><th>URI:</th><td>/pluginManager/plugin</td></tr>
<tr><th>STATUS:</th><td>404</td></tr>
<tr><th>MESSAGE:</th><td>Not Found</td></tr>
<tr><th>SERVLET:</th><td>Stapler</td></tr>
</table>
<hr><a href="https://eclipse.org/jetty">Powered by Jetty:// 9.4.43.v20210629</a><hr/>
</body>
</html>
#Task from Ansible playbook:
- name: Running through and installing specified plugins
jenkins_plugin:
name: '{{ item.key }}'
version: '{{ item.value }}'
jenkins_home: '{{ jenkins_home_dir }}'
url_username: 'admin'
url_password: '{{ jenkins_administrator_password }}'
state: pinned
timeout: 120
url: 'http://{{ jenkins_fqdn }}:{{ jenkins_http_port }}'
with_dict:
- "{{ jenkins_install_plugins| items2dict(key_name='name', value_name='version') }}"
notify: Restart Jenkins service
register: plugin_install_result
until: plugin_install_result is success
retries: 3
delay: 10
```
### Expected Results
h-4.2$ curl http://localhost:8080/pluginManager/plugins | jq '.data[]|select(.name == "http_request")'
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
100 906k 100 906k 0 0 3543k 0 --:--:-- --:--:-- --:--:-- 3567k
{
"installed": false,
"name": "http_request",
"title": "HTTP Request",
"excerpt": "This plugin sends a http request to an url with some parameters",
"site": "default",
"dependencies": {
"javax-mail-api": "1.6.2-5",
"workflow-durable-task-step": "2.40",
"apache-httpcomponents-client-4-api": "4.5.13-1.0",
"credentials": "2.6.1",
"javax-activation-api": "1.2.0-2",
"script-security": "1.78",
"sshd": "3.0.1",
"workflow-basic-steps": "2.22"
},
"website": "https://plugins.jenkins.io/http_request"
}
### Actual Results
```console (paste below)
sh-4.2$ curl http://localhost:8080/pluginManager/plugin | jq '.data[]|select(.name == "http_request")'
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
100 464 100 464 0 0 11397 0 --:--:-- --:--:-- --:--:-- 11600
parse error: Invalid numeric literal at line 2, column 0
```
### Code of Conduct
- [X] I agree to follow the Ansible Code of Conduct | 1.0 | Fix wrong url path in plugins/modules/web_infrastructure/jenkins_plugin.py - ### Summary
When we use option from community.general.jenkins_plugin module to verify installed plugins. We received error like below:
`failed: [localhost] (item={'key': 'http_request', 'value': '1.15'}) => {"ansible_loop_var": "item", "attempts": 3, "changed": false, "details": "HTTP Error 404: Not Found", "item": {"key": "http_request", "value": "1.15"}, "msg": "Plugin not found. http://localhost:8080/pluginManager/plugin/http_request/pin"}`
There is wrong url path "pluginManager/plugin" **Which need to be modified to** "pluginManager/plugins"
Follow link below:
https://github.com/ansible-collections/community.general/blob/788cfb624aa7e694b533b5d2722aab0ff4522ce6/plugins/modules/web_infrastructure/jenkins_plugin.py#L774
### Issue Type
Bug Report
### Component Name
plugins/modules/web_infrastructure/jenkins_plugin.py
### Ansible Version
```console (paste below)
sh-4.2$ ansible --version
[DEPRECATION WARNING]: Ansible will require Python 3.8 or newer on the controller starting with Ansible 2.12. Current version: 3.7.10 (default, Jun 3 2021, 00:02:01) [GCC 7.3.1
20180712 (Red Hat 7.3.1-13)]. This feature will be removed from ansible-core in version 2.12. Deprecation warnings can be disabled by setting deprecation_warnings=False in
ansible.cfg.
ansible [core 2.11.12]
config file = /etc/ansible/ansible.cfg
configured module search path = ['/home/ssm-user/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules']
ansible python module location = /usr/local/lib/python3.7/site-packages/ansible
ansible collection location = /home/ssm-user/.ansible/collections:/usr/share/ansible/collections
executable location = /usr/local/bin/ansible
python version = 3.7.10 (default, Jun 3 2021, 00:02:01) [GCC 7.3.1 20180712 (Red Hat 7.3.1-13)]
jinja version = 3.1.2
libyaml = True
```
### Community.general Version
```console (paste below)
$ ansible-galaxy collection list community.general
# /usr/local/lib/python3.7/site-packages/ansible_collections
Collection Version
----------------- -------
community.general 3.8.3
```
### Configuration
```console (paste below)
$ ansible-config dump --only-changed
```
### OS / Environment
_No response_
### Steps to Reproduce
<!--- Paste example playbooks or commands between quotes below -->
```yaml (paste below)
#Simple http request on Jenkins controller
sh-4.2$ curl http://localhost:8080/pluginManager/plugin
<html>
<head>
<meta http-equiv="Content-Type" content="text/html;charset=utf-8"/>
<title>Error 404 Not Found</title>
</head>
<body><h2>HTTP ERROR 404 Not Found</h2>
<table>
<tr><th>URI:</th><td>/pluginManager/plugin</td></tr>
<tr><th>STATUS:</th><td>404</td></tr>
<tr><th>MESSAGE:</th><td>Not Found</td></tr>
<tr><th>SERVLET:</th><td>Stapler</td></tr>
</table>
<hr><a href="https://eclipse.org/jetty">Powered by Jetty:// 9.4.43.v20210629</a><hr/>
</body>
</html>
#Task from Ansible playbook:
- name: Running through and installing specified plugins
jenkins_plugin:
name: '{{ item.key }}'
version: '{{ item.value }}'
jenkins_home: '{{ jenkins_home_dir }}'
url_username: 'admin'
url_password: '{{ jenkins_administrator_password }}'
state: pinned
timeout: 120
url: 'http://{{ jenkins_fqdn }}:{{ jenkins_http_port }}'
with_dict:
- "{{ jenkins_install_plugins| items2dict(key_name='name', value_name='version') }}"
notify: Restart Jenkins service
register: plugin_install_result
until: plugin_install_result is success
retries: 3
delay: 10
```
### Expected Results
h-4.2$ curl http://localhost:8080/pluginManager/plugins | jq '.data[]|select(.name == "http_request")'
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
100 906k 100 906k 0 0 3543k 0 --:--:-- --:--:-- --:--:-- 3567k
{
"installed": false,
"name": "http_request",
"title": "HTTP Request",
"excerpt": "This plugin sends a http request to an url with some parameters",
"site": "default",
"dependencies": {
"javax-mail-api": "1.6.2-5",
"workflow-durable-task-step": "2.40",
"apache-httpcomponents-client-4-api": "4.5.13-1.0",
"credentials": "2.6.1",
"javax-activation-api": "1.2.0-2",
"script-security": "1.78",
"sshd": "3.0.1",
"workflow-basic-steps": "2.22"
},
"website": "https://plugins.jenkins.io/http_request"
}
### Actual Results
```console (paste below)
sh-4.2$ curl http://localhost:8080/pluginManager/plugin | jq '.data[]|select(.name == "http_request")'
% Total % Received % Xferd Average Speed Time Time Time Current
Dload Upload Total Spent Left Speed
100 464 100 464 0 0 11397 0 --:--:-- --:--:-- --:--:-- 11600
parse error: Invalid numeric literal at line 2, column 0
```
### Code of Conduct
- [X] I agree to follow the Ansible Code of Conduct | infrastructure | fix wrong url path in plugins modules web infrastructure jenkins plugin py summary when we use option from community general jenkins plugin module to verify installed plugins we received error like below failed item key http request value ansible loop var item attempts changed false details http error not found item key http request value msg plugin not found there is wrong url path pluginmanager plugin which need to be modified to pluginmanager plugins follow link below issue type bug report component name plugins modules web infrastructure jenkins plugin py ansible version console paste below sh ansible version ansible will require python or newer on the controller starting with ansible current version default jun gcc red hat this feature will be removed from ansible core in version deprecation warnings can be disabled by setting deprecation warnings false in ansible cfg ansible config file etc ansible ansible cfg configured module search path ansible python module location usr local lib site packages ansible ansible collection location home ssm user ansible collections usr share ansible collections executable location usr local bin ansible python version default jun jinja version libyaml true community general version console paste below ansible galaxy collection list community general usr local lib site packages ansible collections collection version community general configuration console paste below ansible config dump only changed os environment no response steps to reproduce yaml paste below simple http request on jenkins controller sh curl error not found http error not found uri pluginmanager plugin status message not found servlet stapler task from ansible playbook name running through and installing specified plugins jenkins plugin name item key version item value jenkins home jenkins home dir url username admin url password jenkins administrator password state pinned timeout url jenkins fqdn jenkins http port with dict jenkins install plugins key name name value name version notify restart jenkins service register plugin install result until plugin install result is success retries delay expected results h curl jq data select name http request total received xferd average speed time time time current dload upload total spent left speed installed false name http request title http request excerpt this plugin sends a http request to an url with some parameters site default dependencies javax mail api workflow durable task step apache httpcomponents client api credentials javax activation api script security sshd workflow basic steps website actual results console paste below sh curl jq data select name http request total received xferd average speed time time time current dload upload total spent left speed parse error invalid numeric literal at line column code of conduct i agree to follow the ansible code of conduct | 1 |
3,349 | 4,241,079,086 | IssuesEvent | 2016-07-06 15:18:24 | trackpete/exiletools-indexer | https://api.github.com/repos/trackpete/exiletools-indexer | closed | SSL Cert Expired | infrastructure / exiletools.com setup | The certificate for https://api.exiletools.com/ expired on June 30th. Looks like it was issued by Let's Encrypt. They have a lot of tools that help you automate cert renewal without human interaction that would help keep it up-to-date. | 1.0 | SSL Cert Expired - The certificate for https://api.exiletools.com/ expired on June 30th. Looks like it was issued by Let's Encrypt. They have a lot of tools that help you automate cert renewal without human interaction that would help keep it up-to-date. | infrastructure | ssl cert expired the certificate for expired on june looks like it was issued by let s encrypt they have a lot of tools that help you automate cert renewal without human interaction that would help keep it up to date | 1 |
32,473 | 26,722,382,023 | IssuesEvent | 2023-01-29 09:43:30 | rusefi/rusefi | https://api.github.com/repos/rusefi/rusefi | closed | GHA: submodules: recursive | INFRASTRUCTURE | Do we really need ``submodules: recursive`` or can we save the planet by not?

| 1.0 | GHA: submodules: recursive - Do we really need ``submodules: recursive`` or can we save the planet by not?

| infrastructure | gha submodules recursive do we really need submodules recursive or can we save the planet by not | 1 |
602,432 | 18,468,996,423 | IssuesEvent | 2021-10-17 12:06:39 | python-discord/site | https://api.github.com/repos/python-discord/site | closed | Remove Subdomains From Bot Guide | area: docs priority: 0 - critical status: stalled type: enhancement | #526 and python-discord/bot#1630 removed subdomains from the bot and site this week, but we still need to update the guides to reflect those changes.
What needs to be done:
- Under the [configure the bot](https://www.pythondiscord.com/pages/guides/pydis-guides/contributing/bot/#configure-the-bot) section, we will need to remove the mentions of subdomains, and instead mention query paths (`localhost:8000/api` for instance).
- We should delete the guide on [setting up hosts](https://www.pythondiscord.com/pages/guides/pydis-guides/contributing/hosts-file/).
> ❗ Work on this does not need to start until #591 is merged to avoid conflicts | 1.0 | Remove Subdomains From Bot Guide - #526 and python-discord/bot#1630 removed subdomains from the bot and site this week, but we still need to update the guides to reflect those changes.
What needs to be done:
- Under the [configure the bot](https://www.pythondiscord.com/pages/guides/pydis-guides/contributing/bot/#configure-the-bot) section, we will need to remove the mentions of subdomains, and instead mention query paths (`localhost:8000/api` for instance).
- We should delete the guide on [setting up hosts](https://www.pythondiscord.com/pages/guides/pydis-guides/contributing/hosts-file/).
> ❗ Work on this does not need to start until #591 is merged to avoid conflicts | non_infrastructure | remove subdomains from bot guide and python discord bot removed subdomains from the bot and site this week but we still need to update the guides to reflect those changes what needs to be done under the section we will need to remove the mentions of subdomains and instead mention query paths localhost api for instance we should delete the guide on ❗ work on this does not need to start until is merged to avoid conflicts | 0 |
10,513 | 8,603,207,162 | IssuesEvent | 2018-11-16 16:12:19 | firebase/firebase-ios-sdk | https://api.github.com/repos/firebase/firebase-ios-sdk | closed | Run travis pod spec lint testing for Xcode 9 | Infrastructure | The release 5.13.0 release testing discovered two Xcode 9 pod spec lint errors that don't show up in Xcode 10. We need to add Xcode 9 lint checking to regular Travis regression testing.
- #2079 fixed introduced nullability warnings in GoogleUtilities
- #2032 introduced Xcode 9 build errors for InAppMessaging
cc: @Corrob | 1.0 | Run travis pod spec lint testing for Xcode 9 - The release 5.13.0 release testing discovered two Xcode 9 pod spec lint errors that don't show up in Xcode 10. We need to add Xcode 9 lint checking to regular Travis regression testing.
- #2079 fixed introduced nullability warnings in GoogleUtilities
- #2032 introduced Xcode 9 build errors for InAppMessaging
cc: @Corrob | infrastructure | run travis pod spec lint testing for xcode the release release testing discovered two xcode pod spec lint errors that don t show up in xcode we need to add xcode lint checking to regular travis regression testing fixed introduced nullability warnings in googleutilities introduced xcode build errors for inappmessaging cc corrob | 1 |
1,833 | 6,802,560,923 | IssuesEvent | 2017-11-02 20:39:07 | vmware/vic | https://api.github.com/repos/vmware/vic | closed | Refine patterns used for vic-machine service end-to-end tests | area/apis component/test component/vic-machine kind/architecture team/lifecycle | The vic-machine service product is establishing new patterns for organization of end-to-end test code. Towards the end of the project, we should revisit the results and make sure it's what we want to merge to `master`.
See https://github.com/vmware/vic/pull/6475#pullrequestreview-66940252 for more context. | 1.0 | Refine patterns used for vic-machine service end-to-end tests - The vic-machine service product is establishing new patterns for organization of end-to-end test code. Towards the end of the project, we should revisit the results and make sure it's what we want to merge to `master`.
See https://github.com/vmware/vic/pull/6475#pullrequestreview-66940252 for more context. | non_infrastructure | refine patterns used for vic machine service end to end tests the vic machine service product is establishing new patterns for organization of end to end test code towards the end of the project we should revisit the results and make sure it s what we want to merge to master see for more context | 0 |
330,712 | 24,274,115,799 | IssuesEvent | 2022-09-28 12:41:29 | eheh12321/InterviewHelper | https://api.github.com/repos/eheh12321/InterviewHelper | closed | 도메인 수정 | documentation enhancement | #6 과 관련해서 수정이 필요한 도메인을 수정하고자 함
[수정 예정 사항]
* [x] `UserAccount` 도메인에서 AuditingFields 삭제
- 회원 계정은 본인만 수정 가능한데 CreatedBy와 같은 필드는 의미가 없다고 생각함.
* [x] `@Setter` 제거
- 변경은 별도의 메서드를 이용해서만 변경될 수 있도록.
* [x] `@Column` 속성에 들어가는 length, nullable과 같은 속성 ERD에 업데이트
* [x] public으로 열려있는 생성자 private으로 닫고 팩토리 메서드 (.of()) 통해 생성할 수 있도록 수정 | 1.0 | 도메인 수정 - #6 과 관련해서 수정이 필요한 도메인을 수정하고자 함
[수정 예정 사항]
* [x] `UserAccount` 도메인에서 AuditingFields 삭제
- 회원 계정은 본인만 수정 가능한데 CreatedBy와 같은 필드는 의미가 없다고 생각함.
* [x] `@Setter` 제거
- 변경은 별도의 메서드를 이용해서만 변경될 수 있도록.
* [x] `@Column` 속성에 들어가는 length, nullable과 같은 속성 ERD에 업데이트
* [x] public으로 열려있는 생성자 private으로 닫고 팩토리 메서드 (.of()) 통해 생성할 수 있도록 수정 | non_infrastructure | 도메인 수정 과 관련해서 수정이 필요한 도메인을 수정하고자 함 useraccount 도메인에서 auditingfields 삭제 회원 계정은 본인만 수정 가능한데 createdby와 같은 필드는 의미가 없다고 생각함 setter 제거 변경은 별도의 메서드를 이용해서만 변경될 수 있도록 column 속성에 들어가는 length nullable과 같은 속성 erd에 업데이트 public으로 열려있는 생성자 private으로 닫고 팩토리 메서드 of 통해 생성할 수 있도록 수정 | 0 |
129,055 | 5,087,901,797 | IssuesEvent | 2016-12-31 11:35:08 | ITDevLtd/MCVirt | https://api.github.com/repos/ITDevLtd/MCVirt | closed | Handle in-use ports during DRBD creation | High Priority | If a port is in use (at at least DRBD thinks it is), the creation will fail with:
2016-11-25 09:11:29,027 mcvirtd /usr/lib/mcvirt/system.py 42: runCommand ERROR Failed system command: /sbin/drbdadm, connect, mcvirt_vm-VMNAME-disk-1
RC: 10
Stdout:
Stderr: mcvirt_vm-VMNAME-disk-1: Failure: (102) Local address(port) already in use.
Command 'drbdsetup-84 connect mcvirt_vm-VMNAME-disk-1 ipv4:***:*** ipv4:***:*** --verify-alg=sha1 --after-sb-2pri=disconnect --after-sb-1pri=discard-secondary --after-sb-0pri=discard-zero-changes --shared-secret=*** --cram-hmac-alg=sha1 --ko-count=50 --protocol=C' terminated with exit code 10
2016-11-25 09:11:31,022 mcvirtd /usr/lib/mcvirt/rpc/lock.py 83: lock_log_and_call ERROR An internal MCVirt exception occurred in lock
2016-11-25 09:11:31,024 mcvirtd /usr/lib/mcvirt/rpc/lock.py 84: lock_log_and_call ERROR Traceback (most recent call last):
File "/usr/lib/mcvirt/rpc/lock.py", line 81, in lock_log_and_call
response = callback(*args, **kwargs)
File "/usr/lib/mcvirt/virtual_machine/virtual_machine.py", line 1063, in duplicate
disk_object.duplicate(new_vm_object)
File "/usr/lib/mcvirt/virtual_machine/hard_drive/base.py", line 205, in duplicate
new_disk_object.create(self.getSize())
File "/usr/lib/mcvirt/virtual_machine/hard_drive/drbd.py", line 369, in create
self._drbdConnect()
File "/usr/lib/mcvirt/virtual_machine/hard_drive/drbd.py", line 581, in _drbdConnect
System.runCommand([NodeDrbd.DrbdADM, 'connect', self.resource_name])
File "/usr/lib/mcvirt/system.py", line 43, in runCommand
raise MCVirtCommandException(('External command failure. '
MCVirtCommandException: External command failure. See MCVirt log for more information
| 1.0 | Handle in-use ports during DRBD creation - If a port is in use (at at least DRBD thinks it is), the creation will fail with:
2016-11-25 09:11:29,027 mcvirtd /usr/lib/mcvirt/system.py 42: runCommand ERROR Failed system command: /sbin/drbdadm, connect, mcvirt_vm-VMNAME-disk-1
RC: 10
Stdout:
Stderr: mcvirt_vm-VMNAME-disk-1: Failure: (102) Local address(port) already in use.
Command 'drbdsetup-84 connect mcvirt_vm-VMNAME-disk-1 ipv4:***:*** ipv4:***:*** --verify-alg=sha1 --after-sb-2pri=disconnect --after-sb-1pri=discard-secondary --after-sb-0pri=discard-zero-changes --shared-secret=*** --cram-hmac-alg=sha1 --ko-count=50 --protocol=C' terminated with exit code 10
2016-11-25 09:11:31,022 mcvirtd /usr/lib/mcvirt/rpc/lock.py 83: lock_log_and_call ERROR An internal MCVirt exception occurred in lock
2016-11-25 09:11:31,024 mcvirtd /usr/lib/mcvirt/rpc/lock.py 84: lock_log_and_call ERROR Traceback (most recent call last):
File "/usr/lib/mcvirt/rpc/lock.py", line 81, in lock_log_and_call
response = callback(*args, **kwargs)
File "/usr/lib/mcvirt/virtual_machine/virtual_machine.py", line 1063, in duplicate
disk_object.duplicate(new_vm_object)
File "/usr/lib/mcvirt/virtual_machine/hard_drive/base.py", line 205, in duplicate
new_disk_object.create(self.getSize())
File "/usr/lib/mcvirt/virtual_machine/hard_drive/drbd.py", line 369, in create
self._drbdConnect()
File "/usr/lib/mcvirt/virtual_machine/hard_drive/drbd.py", line 581, in _drbdConnect
System.runCommand([NodeDrbd.DrbdADM, 'connect', self.resource_name])
File "/usr/lib/mcvirt/system.py", line 43, in runCommand
raise MCVirtCommandException(('External command failure. '
MCVirtCommandException: External command failure. See MCVirt log for more information
| non_infrastructure | handle in use ports during drbd creation if a port is in use at at least drbd thinks it is the creation will fail with mcvirtd usr lib mcvirt system py runcommand error failed system command sbin drbdadm connect mcvirt vm vmname disk rc stdout stderr mcvirt vm vmname disk failure local address port already in use command drbdsetup connect mcvirt vm vmname disk verify alg after sb disconnect after sb discard secondary after sb discard zero changes shared secret cram hmac alg ko count protocol c terminated with exit code mcvirtd usr lib mcvirt rpc lock py lock log and call error an internal mcvirt exception occurred in lock mcvirtd usr lib mcvirt rpc lock py lock log and call error traceback most recent call last file usr lib mcvirt rpc lock py line in lock log and call response callback args kwargs file usr lib mcvirt virtual machine virtual machine py line in duplicate disk object duplicate new vm object file usr lib mcvirt virtual machine hard drive base py line in duplicate new disk object create self getsize file usr lib mcvirt virtual machine hard drive drbd py line in create self drbdconnect file usr lib mcvirt virtual machine hard drive drbd py line in drbdconnect system runcommand file usr lib mcvirt system py line in runcommand raise mcvirtcommandexception external command failure mcvirtcommandexception external command failure see mcvirt log for more information | 0 |
25,546 | 18,846,824,742 | IssuesEvent | 2021-11-11 15:48:56 | pytest-dev/pytest | https://api.github.com/repos/pytest-dev/pytest | closed | Figure out how to fix changelog title | type: infrastructure | Strange, this should be:
```suggestion
pytest 7.0.0rc1 (2021-06-28)
============================
```
No idea, seems like a town crier problem? 🤔
_Originally posted by @nicoddemus in https://github.com/pytest-dev/pytest/pull/8814#discussion_r660066152_ | 1.0 | Figure out how to fix changelog title - Strange, this should be:
```suggestion
pytest 7.0.0rc1 (2021-06-28)
============================
```
No idea, seems like a town crier problem? 🤔
_Originally posted by @nicoddemus in https://github.com/pytest-dev/pytest/pull/8814#discussion_r660066152_ | infrastructure | figure out how to fix changelog title strange this should be suggestion pytest no idea seems like a town crier problem 🤔 originally posted by nicoddemus in | 1 |
16,459 | 11,966,159,800 | IssuesEvent | 2020-04-06 02:24:35 | department-of-veterans-affairs/va.gov-cms | https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms | closed | [cms-ci] Sync files from upstream | DevOps Epic ⭐️ Infrastructure | Sync files from upstream and have all new environments consume that tar file. Create similar command as `cms-db-download` e.g. `cms-files-download`.
This should be synced at same interval as DB, `10 * * * *`.
Current logs have a lot of "image download failed" in logs.
cc @swirtSJW | 1.0 | [cms-ci] Sync files from upstream - Sync files from upstream and have all new environments consume that tar file. Create similar command as `cms-db-download` e.g. `cms-files-download`.
This should be synced at same interval as DB, `10 * * * *`.
Current logs have a lot of "image download failed" in logs.
cc @swirtSJW | infrastructure | sync files from upstream sync files from upstream and have all new environments consume that tar file create similar command as cms db download e g cms files download this should be synced at same interval as db current logs have a lot of image download failed in logs cc swirtsjw | 1 |
8,673 | 7,551,122,256 | IssuesEvent | 2018-04-18 19:04:43 | dotnet/cli | https://api.github.com/repos/dotnet/cli | closed | cli-test-env.sh cannot run test on macOS | infrastructure | There is 2 problem, `uname` is not evaluated, the following can make it work. However the hard part is RID in mac has the version number. Which it should not
```sh
uname=$(uname)
if [ "$(uname)" == "Darwin" ]
then
RID=osx.10.13-x64
else
RID=linux-x64
fi
```
| 1.0 | cli-test-env.sh cannot run test on macOS - There is 2 problem, `uname` is not evaluated, the following can make it work. However the hard part is RID in mac has the version number. Which it should not
```sh
uname=$(uname)
if [ "$(uname)" == "Darwin" ]
then
RID=osx.10.13-x64
else
RID=linux-x64
fi
```
| infrastructure | cli test env sh cannot run test on macos there is problem uname is not evaluated the following can make it work however the hard part is rid in mac has the version number which it should not sh uname uname if then rid osx else rid linux fi | 1 |
21,524 | 14,617,441,419 | IssuesEvent | 2020-12-22 14:47:35 | ionide/ionide-vscode-fsharp | https://api.github.com/repos/ionide/ionide-vscode-fsharp | closed | add smoke tests for ionide extension | enhancement infrastructure up for grabs | ref http://vscode-docs.readthedocs.io/en/stable/extensions/testing-extensions/
ref https://code.visualstudio.com/docs/extensions/testing-extensions
to add basic smoke test for os, in travis/appveyor
| 1.0 | add smoke tests for ionide extension - ref http://vscode-docs.readthedocs.io/en/stable/extensions/testing-extensions/
ref https://code.visualstudio.com/docs/extensions/testing-extensions
to add basic smoke test for os, in travis/appveyor
| infrastructure | add smoke tests for ionide extension ref ref to add basic smoke test for os in travis appveyor | 1 |
32,510 | 26,747,253,844 | IssuesEvent | 2023-01-30 16:49:07 | crossplane/docs | https://api.github.com/repos/crossplane/docs | closed | Sanitize expand inputs | bug P2 infrastructure | The label used in the `expand` shortcode becomes a javascript selector. If the input isn't valid, like a period, then the expand box just doesn't work.
The shortcode should sanitize the input instead of forcing users to change what they write. | 1.0 | Sanitize expand inputs - The label used in the `expand` shortcode becomes a javascript selector. If the input isn't valid, like a period, then the expand box just doesn't work.
The shortcode should sanitize the input instead of forcing users to change what they write. | infrastructure | sanitize expand inputs the label used in the expand shortcode becomes a javascript selector if the input isn t valid like a period then the expand box just doesn t work the shortcode should sanitize the input instead of forcing users to change what they write | 1 |
6,028 | 6,114,248,252 | IssuesEvent | 2017-06-22 00:22:30 | redox-os/ion | https://api.github.com/repos/redox-os/ion | closed | Travis should report coverage statistics to coveralls.io | infrastructure | Tracking test coverage would add a more visible impact for writing unit tests. We already have a badge on our README for it, but we don't send any reports.
| 1.0 | Travis should report coverage statistics to coveralls.io - Tracking test coverage would add a more visible impact for writing unit tests. We already have a badge on our README for it, but we don't send any reports.
| infrastructure | travis should report coverage statistics to coveralls io tracking test coverage would add a more visible impact for writing unit tests we already have a badge on our readme for it but we don t send any reports | 1 |
16,919 | 12,152,149,356 | IssuesEvent | 2020-04-24 21:31:19 | BCDevOps/developer-experience | https://api.github.com/repos/BCDevOps/developer-experience | closed | Explore moving gluster volume health checks from serial to parallel process | Infrastructure closed | ERROR: type should be string, got "https://trello.com/c/8jIrNldM/59-explore-moving-gluster-volume-health-checks-from-serial-to-parallel-process\n\nWith CNS volume count making a healthy increase in light of legacy gluster storage going away, we need to look at making gluster volume health checks a parallel process. \nThis step is now 100% the bottleneck during any operation requiring us to perform volume health checks (OCS/OCP upgrades, OS patching). " | 1.0 | Explore moving gluster volume health checks from serial to parallel process - https://trello.com/c/8jIrNldM/59-explore-moving-gluster-volume-health-checks-from-serial-to-parallel-process
With CNS volume count making a healthy increase in light of legacy gluster storage going away, we need to look at making gluster volume health checks a parallel process.
This step is now 100% the bottleneck during any operation requiring us to perform volume health checks (OCS/OCP upgrades, OS patching). | infrastructure | explore moving gluster volume health checks from serial to parallel process with cns volume count making a healthy increase in light of legacy gluster storage going away we need to look at making gluster volume health checks a parallel process this step is now the bottleneck during any operation requiring us to perform volume health checks ocs ocp upgrades os patching | 1 |
204,669 | 23,270,307,402 | IssuesEvent | 2022-08-04 22:04:15 | MatBenfield/news | https://api.github.com/repos/MatBenfield/news | closed | [SecurityWeek] Compliance Automation Startup RegScale Scores $20 Million Investment | SecurityWeek Stale |
**RegScale, a Virginia startup building technology to manage continuous compliance automation tasks, has attracted $20 million in early-stage venture capital funding.**
The Series A round was led by SYN Ventures with participation from SineWave Ventures, VIPC’s Virginia Venture Partners and SecureOctane.
[read more](https://www.securityweek.com/compliance-automation-startup-regscale-scores-20-million-investment)
<https://www.securityweek.com/compliance-automation-startup-regscale-scores-20-million-investment>
| True | [SecurityWeek] Compliance Automation Startup RegScale Scores $20 Million Investment -
**RegScale, a Virginia startup building technology to manage continuous compliance automation tasks, has attracted $20 million in early-stage venture capital funding.**
The Series A round was led by SYN Ventures with participation from SineWave Ventures, VIPC’s Virginia Venture Partners and SecureOctane.
[read more](https://www.securityweek.com/compliance-automation-startup-regscale-scores-20-million-investment)
<https://www.securityweek.com/compliance-automation-startup-regscale-scores-20-million-investment>
| non_infrastructure | compliance automation startup regscale scores million investment regscale a virginia startup building technology to manage continuous compliance automation tasks has attracted million in early stage venture capital funding the series a round was led by syn ventures with participation from sinewave ventures vipc’s virginia venture partners and secureoctane | 0 |
29,684 | 24,171,922,990 | IssuesEvent | 2022-09-22 20:05:05 | oppia/oppia-android | https://api.github.com/repos/oppia/oppia-android | closed | Develop branch is broken | issue_user_impact_high issue_type_infrastructure issue_temp_ben_triaged issue_user_developer | **Describe the bug**
Two PRs were merged in succession that led to develop breaking, despite no conflicts.
**To Reproduce**
See develop branch.
**Expected behavior**
Develop shouldn't be broken.
**Demonstration**
N/A
**Environment**
N/A
**Additional context**
N/A
| 1.0 | Develop branch is broken - **Describe the bug**
Two PRs were merged in succession that led to develop breaking, despite no conflicts.
**To Reproduce**
See develop branch.
**Expected behavior**
Develop shouldn't be broken.
**Demonstration**
N/A
**Environment**
N/A
**Additional context**
N/A
| infrastructure | develop branch is broken describe the bug two prs were merged in succession that led to develop breaking despite no conflicts to reproduce see develop branch expected behavior develop shouldn t be broken demonstration n a environment n a additional context n a | 1 |
7,716 | 7,058,366,787 | IssuesEvent | 2018-01-04 20:04:46 | angular/material2 | https://api.github.com/repos/angular/material2 | closed | Deprecation warnings in _theming.scss | has pr infrastructure P3: important | #### Bug, feature request, or proposal:
Bug.
#### What is the expected behavior?
No warnings during build.
#### What is the current behavior?
I see the following self-explaining deprecation warnings:
```
ng build --prod
10% building modules 4/5 modules 1 active ...\<project_root>l\src\theme.scssDEPRECATION WARNING: Passing null, a non-string value, to unquote()
will be an error in future versions of Sass.
on line 1483 of <project_root>/node_modules/@angular/material/_theming.scss
DEPRECATION WARNING: Passing null, a non-string value, to unquote()
will be an error in future versions of Sass.
on line 1483 of c:/Work/Dita 10% building modules 6/6 modules 0 activeDEPRECATION WARNING: Passing null, a non-string value, to unquote()
will be an error in future versions of Sass.
on line 1483 of <project_root>/node_modules/@angular/material/_theming.scss
DEPRECATION WARNING: Passing null, a non-string value, to unquote()
will be an error in future versions of Sass.
on line 1483 of <project_root>/node_modules/@angular/material/_theming.scss
DEPRECATION WARNING: Passing null, a non-string value, to unquote()
will be an error in future versions of Sass.
on line 1483 of <project_root>/node_modules/@angular/material/_theming.scss
DEPRECATION WARNING: Passing null, a non-string value, to unquote()
will be an error in future versions of Sass.
on line 1483 of <project_root>/node_modules/@angular/material/_theming.scss
DEPRECATION WARNING: Passing null, a non-string value, to unquote()
DEPRECATION WARNING: Passing null, a non-string value, to unquote()
will be an error in future versions of Sass.
on line 1483 of <project_root>/node_modules/@angular/material/_theming.scss
will be an error in future versions of Sass.
on line 1483 of <project_root>/node_modules/@angular/material/_theming.scss
DEPRECATION WARNING: Passing null, a non-string value, to unquote()
will be an error in future versions of Sass.
on line 1483 of <project_root>/node_modules/@angular/material/_theming.scss
DEPRECATION WARNING: Passing null, a non-string value, to unquote()
will be an error in future versions of Sass.
on line 1483 of <project_root>/node_modules/@angular 69% building modules 740/741 modules 1 active ...itatWeb\<project_root>l\src\theme.scssDEPRECATION WARNING: Passing null, a non-string value, to unquote()
will be an error in future versions of Sass.
on line 1483 of <project_root>/node_modules/@angular/material/_theming.scss
DEPRECATION WARNING: Passing null, a non-string value, to unquote()
will be an error in future versions of Sass.
on line 1483 of c:/Work/Date: 2017-12-13T20:43:31.018Z
Hash: 0fe6fd8d1d7e608d6e79
Time: 112769ms
chunk {0} polyfills.0c6625b3fc183e5d4767.bundle.js (polyfills) 96.4 kB [initial] [rendered]
chunk {1} main.9b8205e180b9751fc93c.bundle.js (main) 1.47 MB [initial] [rendered]
chunk {2} styles.525ebb3f34a68462be9d.bundle.css (styles) 46 kB [initial] [rendered]
chunk {3} inline.b240d1c2aa548af8fc2a.bundle.js (inline) 1.45 kB [entry] [rendered]
```
#### What are the steps to reproduce?
1. Create a new app with CLI.
2. Add Material packages to `package.json`:
```
"dependencies": {
"@angular/animations": "^5.1.1",
"@angular/cdk": "^5.0.1",
"@angular/common": "^5.1.1",
"@angular/compiler": "^5.1.1",
"@angular/core": "^5.1.1",
"@angular/flex-layout": "2.0.0-beta.10-4905443",
"@angular/forms": "^5.1.1",
"@angular/http": "^5.1.1",
"@angular/material": "^5.0.1",
"@angular/platform-browser": "^5.1.1",
"@angular/platform-browser-dynamic": "^5.1.1",
"@angular/router": "^5.1.1",
"core-js": "^2.5.3",
"rxjs": "^5.5.5",
"zone.js": "^0.8.14"
},
```
3. Create a custom `theme.scss` in `src` folder and register it in `.angular-cli.json`:
```
@import '~@angular/material/theming';
$typography-config-default: mat-typography-config();
$typography-config-custom: (
button: mat-typography-level(18px, 18px, bold) /* This line causes warnings */
);
$typography-config: map-merge($typography-config-default, $typography-config-custom);
@include mat-core($typography-config);
```
4. Run `ng serve` and see the warnings.
#### What is the use-case or motivation for changing an existing behavior?
#### Which versions of Angular, Material, OS, TypeScript, browsers are affected?
Angular 5.1.1, Angular Material 5.0.1, TypeScript 2.4.2.
#### Is there anything else we should know?
| 1.0 | Deprecation warnings in _theming.scss - #### Bug, feature request, or proposal:
Bug.
#### What is the expected behavior?
No warnings during build.
#### What is the current behavior?
I see the following self-explaining deprecation warnings:
```
ng build --prod
10% building modules 4/5 modules 1 active ...\<project_root>l\src\theme.scssDEPRECATION WARNING: Passing null, a non-string value, to unquote()
will be an error in future versions of Sass.
on line 1483 of <project_root>/node_modules/@angular/material/_theming.scss
DEPRECATION WARNING: Passing null, a non-string value, to unquote()
will be an error in future versions of Sass.
on line 1483 of c:/Work/Dita 10% building modules 6/6 modules 0 activeDEPRECATION WARNING: Passing null, a non-string value, to unquote()
will be an error in future versions of Sass.
on line 1483 of <project_root>/node_modules/@angular/material/_theming.scss
DEPRECATION WARNING: Passing null, a non-string value, to unquote()
will be an error in future versions of Sass.
on line 1483 of <project_root>/node_modules/@angular/material/_theming.scss
DEPRECATION WARNING: Passing null, a non-string value, to unquote()
will be an error in future versions of Sass.
on line 1483 of <project_root>/node_modules/@angular/material/_theming.scss
DEPRECATION WARNING: Passing null, a non-string value, to unquote()
will be an error in future versions of Sass.
on line 1483 of <project_root>/node_modules/@angular/material/_theming.scss
DEPRECATION WARNING: Passing null, a non-string value, to unquote()
DEPRECATION WARNING: Passing null, a non-string value, to unquote()
will be an error in future versions of Sass.
on line 1483 of <project_root>/node_modules/@angular/material/_theming.scss
will be an error in future versions of Sass.
on line 1483 of <project_root>/node_modules/@angular/material/_theming.scss
DEPRECATION WARNING: Passing null, a non-string value, to unquote()
will be an error in future versions of Sass.
on line 1483 of <project_root>/node_modules/@angular/material/_theming.scss
DEPRECATION WARNING: Passing null, a non-string value, to unquote()
will be an error in future versions of Sass.
on line 1483 of <project_root>/node_modules/@angular 69% building modules 740/741 modules 1 active ...itatWeb\<project_root>l\src\theme.scssDEPRECATION WARNING: Passing null, a non-string value, to unquote()
will be an error in future versions of Sass.
on line 1483 of <project_root>/node_modules/@angular/material/_theming.scss
DEPRECATION WARNING: Passing null, a non-string value, to unquote()
will be an error in future versions of Sass.
on line 1483 of c:/Work/Date: 2017-12-13T20:43:31.018Z
Hash: 0fe6fd8d1d7e608d6e79
Time: 112769ms
chunk {0} polyfills.0c6625b3fc183e5d4767.bundle.js (polyfills) 96.4 kB [initial] [rendered]
chunk {1} main.9b8205e180b9751fc93c.bundle.js (main) 1.47 MB [initial] [rendered]
chunk {2} styles.525ebb3f34a68462be9d.bundle.css (styles) 46 kB [initial] [rendered]
chunk {3} inline.b240d1c2aa548af8fc2a.bundle.js (inline) 1.45 kB [entry] [rendered]
```
#### What are the steps to reproduce?
1. Create a new app with CLI.
2. Add Material packages to `package.json`:
```
"dependencies": {
"@angular/animations": "^5.1.1",
"@angular/cdk": "^5.0.1",
"@angular/common": "^5.1.1",
"@angular/compiler": "^5.1.1",
"@angular/core": "^5.1.1",
"@angular/flex-layout": "2.0.0-beta.10-4905443",
"@angular/forms": "^5.1.1",
"@angular/http": "^5.1.1",
"@angular/material": "^5.0.1",
"@angular/platform-browser": "^5.1.1",
"@angular/platform-browser-dynamic": "^5.1.1",
"@angular/router": "^5.1.1",
"core-js": "^2.5.3",
"rxjs": "^5.5.5",
"zone.js": "^0.8.14"
},
```
3. Create a custom `theme.scss` in `src` folder and register it in `.angular-cli.json`:
```
@import '~@angular/material/theming';
$typography-config-default: mat-typography-config();
$typography-config-custom: (
button: mat-typography-level(18px, 18px, bold) /* This line causes warnings */
);
$typography-config: map-merge($typography-config-default, $typography-config-custom);
@include mat-core($typography-config);
```
4. Run `ng serve` and see the warnings.
#### What is the use-case or motivation for changing an existing behavior?
#### Which versions of Angular, Material, OS, TypeScript, browsers are affected?
Angular 5.1.1, Angular Material 5.0.1, TypeScript 2.4.2.
#### Is there anything else we should know?
| infrastructure | deprecation warnings in theming scss bug feature request or proposal bug what is the expected behavior no warnings during build what is the current behavior i see the following self explaining deprecation warnings ng build prod building modules modules active l src theme scssdeprecation warning passing null a non string value to unquote will be an error in future versions of sass on line of node modules angular material theming scss deprecation warning passing null a non string value to unquote will be an error in future versions of sass on line of c work dita building modules modules activedeprecation warning passing null a non string value to unquote will be an error in future versions of sass on line of node modules angular material theming scss deprecation warning passing null a non string value to unquote will be an error in future versions of sass on line of node modules angular material theming scss deprecation warning passing null a non string value to unquote will be an error in future versions of sass on line of node modules angular material theming scss deprecation warning passing null a non string value to unquote will be an error in future versions of sass on line of node modules angular material theming scss deprecation warning passing null a non string value to unquote deprecation warning passing null a non string value to unquote will be an error in future versions of sass on line of node modules angular material theming scss will be an error in future versions of sass on line of node modules angular material theming scss deprecation warning passing null a non string value to unquote will be an error in future versions of sass on line of node modules angular material theming scss deprecation warning passing null a non string value to unquote will be an error in future versions of sass on line of node modules angular building modules modules active itatweb l src theme scssdeprecation warning passing null a non string value to unquote will be an error in future versions of sass on line of node modules angular material theming scss deprecation warning passing null a non string value to unquote will be an error in future versions of sass on line of c work date hash time chunk polyfills bundle js polyfills kb chunk main bundle js main mb chunk styles bundle css styles kb chunk inline bundle js inline kb what are the steps to reproduce create a new app with cli add material packages to package json dependencies angular animations angular cdk angular common angular compiler angular core angular flex layout beta angular forms angular http angular material angular platform browser angular platform browser dynamic angular router core js rxjs zone js create a custom theme scss in src folder and register it in angular cli json import angular material theming typography config default mat typography config typography config custom button mat typography level bold this line causes warnings typography config map merge typography config default typography config custom include mat core typography config run ng serve and see the warnings what is the use case or motivation for changing an existing behavior which versions of angular material os typescript browsers are affected angular angular material typescript is there anything else we should know | 1 |
13,147 | 10,131,772,968 | IssuesEvent | 2019-08-01 20:26:00 | HumanCellAtlas/secondary-analysis | https://api.github.com/repos/HumanCellAtlas/secondary-analysis | closed | Audit database permissions | infrastructure | To ensure least privilege principles are followed.
Done for this ticket will be a report of
* permissions on tablespaces
* list of credentials/identities used as well as their age
* remediation plan for updating any credentials older than a year and reconfiguring tablespace permissions to follow least privilege principles
┆Issue is synchronized with this [Jira Story](https://broadinstitute.atlassian.net/browse/GH-267)
| 1.0 | Audit database permissions - To ensure least privilege principles are followed.
Done for this ticket will be a report of
* permissions on tablespaces
* list of credentials/identities used as well as their age
* remediation plan for updating any credentials older than a year and reconfiguring tablespace permissions to follow least privilege principles
┆Issue is synchronized with this [Jira Story](https://broadinstitute.atlassian.net/browse/GH-267)
| infrastructure | audit database permissions to ensure least privilege principles are followed done for this ticket will be a report of permissions on tablespaces list of credentials identities used as well as their age remediation plan for updating any credentials older than a year and reconfiguring tablespace permissions to follow least privilege principles ┆issue is synchronized with this | 1 |
274,823 | 20,868,839,509 | IssuesEvent | 2022-03-22 10:00:29 | phukiendienthoaigiare/Set-Qua-Tang-Doanh-Nghiep-Dip-Tet | https://api.github.com/repos/phukiendienthoaigiare/Set-Qua-Tang-Doanh-Nghiep-Dip-Tet | opened | Nên Tặng Quà Gì Cho Khách Hàng? | documentation help wanted good first issue question | <p><strong>Những món quà tặng vừa hữu ích, vừa hợp thời sẽ có thể “đánh động” đến trái tim của khách hàng và hỗ trợ chiến lược Marketing vô cùng hiệu quả đấy!</strong></p>
<p>Tặng quà cho khách hàng là một trong những phương thức kinh doanh vô cùng hiệu quả. Doanh nghiệp không chỉ có được mối quan hệ tốt với khách hàng, thúc đẩy doanh thu mà còn có thể Marketing thương hiệu một cách tự nhiên. Song, để có thể đạt được tính PR thương hiệu và được nhiều khách hàng yêu thích, không phải món quà nào doanhh nghiệp cũng có thể “tặng đại” mà cần chú trọng vào một số tiêu chí nhất định.</p>
<h2 id="h-cần-quan-tam-gi-khi-chọn-qua-tặng-cho-khach-hang">Cần quan tâm gì khi chọn quà tặng cho khách hàng?</h2>
<p>Có thể nói, việc tặng quà cho khách hàng đã trở thành một chiến lược kinh doanh không thể thiếu trong hầu hết các doanh nghiệp hiện nay. Tùy thuộc vào ngân sách cũng như sản phẩm/dịch vụ đang cung cấp mà giá trị quà tặng có thể lớn hoặc nhỏ. Tuy nhiên, để chiến lược kinh doanh trở nên hiệu quả, có thể duy trì mối quan hệ trong làm ăn (đối với đối tác, khách hàng là doanh nghiệp) hoặc tri ân khách hàng, thúc đẩy tiêu thụ sản phẩm/sử dụng dịch vụ (người tiêu dùng) và thực hiện chiến lược Marketing hiệu quả thì doanh nghiệp nên lựa chọn quà tặng theo các tiêu chí sau:</p>
<h3 id="h-1-chọn-qua-theo-thời-diểm">1. Chọn quà theo thời điểm</h3>
<p>Những món quà tặng theo thời điểm sẽ vừa thể hiện được sự tinh tế, tâm lý và khả năng “đu trend” của doanh nghiệp vừa có thể giúp khách hàng hòa nhập với xu hướng xã hội hiện nay và có thể sử dụng được ngay tại thời điểm đó. Không nhất thiết là những món quà đắt tiền, doanh nghiệp có thể lựa chọn các sản phẩm quà tặng phù hợp với ngân sách của mình, chẳng hạn như áo mưa (tặng vào mùa mưa), quạt mini, bình giữ nhiệt (tặng vào mùa nóng,…),…</p>
<h3 id="h-2-chọn-qua-co-cong-nang-sử-dụng-cao">2. Chọn quà có công năng sử dụng cao</h3>
<p>Ngoại trừ kinh doanh các sản phẩm trang trí hoặc có liên quan đến trang trí, ở những mặt hàng kinh doanh còn lại, doanh nghiệp nên chọn tặng những vật phẩm có công năng sử dụng cao, tần suất sử dụng nhiều và bất kỳ ai cũng có thể dùng được. Điều này sẽ giúp các doanh nghiệp có thể vừa “phủ sóng” thương hiệu của mình vừa mang đến sự hài lòng cho khách hàng, đạt được hiệu quả trong chiến lược kinh doanh.</p>
<h3 id="h-3-chọn-qua-tặng-tương-xứng-với-gia-trị-sản-phẩm">3. Chọn quà tặng tương xứng với giá trị sản phẩm</h3>
<p>Với những khách hàng mua các sản phẩm có giá trị cao, sự mong đợi về món quà tặng cũng tương xứng giá trị và có ít nhiều liên quan đến sản phẩm doanh nghiệp đang cung cấp. Chẳng hạn, nếu doanh nghiệp kinh doanh các sản phẩm điện máy thì có thể lựa chọn các quà tặng là những phụ kiện kèm theo. Điều này sẽ giúp khách hàng cảm thấy mình “lời” khi mua sản phẩm của doanh nghiệp này. Đặc biệt quan trọng, khi những doanh nghiệp tặng kèm quà khuyến mái trong chiến dịch quảng cáo, khách hàng sẽ có thể vì món quà tặng đó mà sẵn sàng móc “hầu bao” để mua sản phẩm.</p>
<p>Trên đây là 3 tiêu chí đóng vai trò là “kim chỉ nam” để doanh nghiệp có thể lựa chọn những món quà tặng phù hợp. Trong số đó, tiêu chí thứ 2 vô cùng quan trọng và được đại đa số doanh nghiệp áp dụng.</p>
<h2 id="h-top-5-qua-tặng-doanh-nghiệp-hợp-thời-dại-2022">Top 5 quà tặng doanh nghiệp hợp thời đại 2022</h2>
<p>Với thời đại công nghệ 4.0 đang phát triển mạnh mẽ như hiện nay thì các doanh nghiệp nên lựa chọn các quà tặng công nghệ phục vụ cho chiến lược kinh doanh của mình. Những món quà tặng công nghệ vừa có tần suất sử dụng cao vừa phù hợp với hầu hết các đối tượng hiện nay, có thể “phủ sóng” thương hiệu một cách tự nhiên nhất.</p>
<h3 id="h-1-loa-bluetooth-khong-day-sieu-mỏng-bỏ-tui-sp005">1. Loa Bluetooth Không Dây Siêu Mỏng Bỏ Túi SP005</h3>
<p>Loa Bluetooth Không Dây Siêu Mỏng SP005 là một dòng loa di động có thiết kế cầm tay vô cùng nhỏ gọn, tiện dụng. Khi tặng món quà này cho khách hàng, họ không chỉ có thể tận hưởng giai điệu bài hát với âm thanh sống động trong thời gian phát nhạc lên đến 3 giờ mà còn có thể mang theo mọi lúc, mọi nơi.</p>
<div class="wp-block-image"><figure class="aligncenter size-full"><img width="800" height="800" src="https://phukiendienthoaigiare.com/wp-content/uploads/2022/02/Loa-Bluetooth-Khong-Day-Sieu-Mong-Bo-Tui-SP005-6.jpg" alt="In logo lên loa làm quà tặng khách hàng độc đáo" class="wp-image-17017" srcset="//phukiendienthoaigiare.com/wp-content/uploads/2022/02/Loa-Bluetooth-Khong-Day-Sieu-Mong-Bo-Tui-SP005-6.jpg 800w, //phukiendienthoaigiare.com/wp-content/uploads/2022/02/Loa-Bluetooth-Khong-Day-Sieu-Mong-Bo-Tui-SP005-6-300x300.jpg 300w, //phukiendienthoaigiare.com/wp-content/uploads/2022/02/Loa-Bluetooth-Khong-Day-Sieu-Mong-Bo-Tui-SP005-6-150x150.jpg 150w, //phukiendienthoaigiare.com/wp-content/uploads/2022/02/Loa-Bluetooth-Khong-Day-Sieu-Mong-Bo-Tui-SP005-6-768x768.jpg 768w, //phukiendienthoaigiare.com/wp-content/uploads/2022/02/Loa-Bluetooth-Khong-Day-Sieu-Mong-Bo-Tui-SP005-6-441x441.jpg 441w, //phukiendienthoaigiare.com/wp-content/uploads/2022/02/Loa-Bluetooth-Khong-Day-Sieu-Mong-Bo-Tui-SP005-6-600x600.jpg 600w" sizes="(max-width: 800px) 100vw, 800px" /><figcaption> Loa Bluetooth Không Dây Siêu Mỏng Bỏ Túi SP005 </figcaption></figure></div>
<p>Được tích hợp công nghệ kết nối không dây Bluetooth 5.0, chiếc loa Bluetooth SP005 này có thể kết nối nhanh chóng với bất kỳ thiết bị điện tử nào (điện thoại, máy tính bảng, laptop,…) và phát ra những âm thanh to, rõ, trong trẻo trong bán kính 10m. Sản phẩm sở hữu công suất 3W cùng dung lượng pin “khủng 4000 mAh nên có thể hoạt động trong thời gian dài, phù hợp sử dụng trong các chuyến du lịch, team-building cùng đồng nghiệp, bạn bè, người thân,…</p>
<p>Loa Bluetooth SP005 đã đạt được chứng nhận an toàn CE/FCC/RoHS nên luôn đảm bảo an toàn cho người sử dụng. Đặc biệt, nếu các doanh nghiệp đặt mua sản phẩm này tại Shop sẽ được miễn phí in/khắc logo lên bề mặt, phục vụ cho chiến lược Marketing và PR thương hiệu.</p>
<h3 id="h-2-tai-nghe-airpods-1-1-huilian-a6-pro">2. Tai Nghe Airpods 1:1 Huilian A6 Pro</h3>
<p>Tai Nghe Airpods 1:1 Huilian A6 Pro được xem như “quà tặng quốc dân” khi được các doanh nghiệp lựa chọn làm quà tặng công nghệ hiện nay. Không chỉ có công năng sử dụng cao mà chi phí đặt mua chiếc tai nghe nhét tai này lại khá thấp, phù hợp với ngân sách của nhiều doanh nghiệp vừa và nhỏ. Hơn thế nữa, những tính năng được tích hợp bên trong chiếc tai nghe này sẽ có thể khiến cho khách hàng của doanh nghiệp cảm thấy ưng ý.</p>
<div class="wp-block-image"><figure class="aligncenter size-full"><img width="600" height="600" src="https://phukiendienthoaigiare.com/wp-content/uploads/2022/03/Nen-Tang-Qua-Gi-Cho-Khach-Hang.png" alt="Tai Nghe Airpods 1:1 Huilian A6 Pro - Quà tặng giá trị cho khách hàng" class="wp-image-18068" srcset="//phukiendienthoaigiare.com/wp-content/uploads/2022/03/Nen-Tang-Qua-Gi-Cho-Khach-Hang.png 600w, //phukiendienthoaigiare.com/wp-content/uploads/2022/03/Nen-Tang-Qua-Gi-Cho-Khach-Hang-300x300.png 300w, //phukiendienthoaigiare.com/wp-content/uploads/2022/03/Nen-Tang-Qua-Gi-Cho-Khach-Hang-150x150.png 150w, //phukiendienthoaigiare.com/wp-content/uploads/2022/03/Nen-Tang-Qua-Gi-Cho-Khach-Hang-441x441.png 441w" sizes="(max-width: 600px) 100vw, 600px" /><figcaption> Tai Nghe Airpods 1:1 Huilian A6 Pro </figcaption></figure></div>
<p>Tai nghe Bluetooth 5.0 Airpods 1:1 Huilian A6 Pro sở hữu thiết kế nhét tai tiên tiến cùng kiểu dáng thời trang, trẻ trung và hiện đại, phù hợp với xu hướng hiện đại hóa và nhịp sống sôi động hiện nay. Do sản phẩm không có dây như kiểu tai nghe nhét tai thông thường nên người dùng có thể điều khiển thông qua cảm ứng được đặt ở bên trong mỗi chiếc tai nghe. Khi sử dụng sản phẩm này, khách hàng của doanh nghiệp có thể nghe nhạc, giải trí, thực hiện cuộc gọi vô cùng dễ dàng với âm thanh trong trẻo, rõ ràng, chất lượng HD ngay cả khi ở những nơi đông người nhờ công nghệ chống ồn tiên tiến. Tính năng này sẽ phù hợp với những ai thường xuyên ra ngoài làm việc hoặc thực hiện các hoạt động ngoài trời.</p>
<p>Tương tự như loa Bluetooth SP005, tai nghe Airpods 1:1 Huilian A6 Pro sở hữu công nghệ 5.0 tiên tiến, không có độ trễ và khả năng chống nhiễu mạnh. Sản phẩm cũng được làm từ chất liệu cao cấp nên khi thực hiện in/khắc logo lên bề mặt sẽ hiển thị những đường nét rõ ràng, hoàn hảo mà không có bất kỳ trầy xước hay hư hỏng gì.</p>
<p><strong>Thông tin sản phẩm:</strong></p>
<ul><li>Kết nối không dây Bluetooth 5.0 không có độ trễ, chống nhiễu mạnh.</li><li>Đàm thoại 2 tai chất lượng cuộc gọi HD cùng công nghệ chống ồn tiên tiến.</li><li>Mở nắp hiện kết nối hiện popup.</li><li>Tính năng đổi tên & định vị tai trong setting điện thoại.</li><li>Hộp sạc hỗ trợ sạc không dây chuẩn Qi, tiện lợi khi sử dụng.</li><li>Bản lề thép chống gập.</li><li>Phù hợp làm quà tặng Doanh Nghiệp, Bạn Bè, Quà khuyến mãi…</li><li>Tùy Chỉnh Logo Doanh nghiệp theo yêu cầu (MOQ=100pcs).</li></ul>
<p><strong>Mua sản phẩm giá rẻ in logo miễn phí cho Doanh nghiệp: <em><a href="https://phukiendienthoaigiare.com/product/tai-nghe-nhet-tai-in-ear-bluetooth-tws-5-0-jacqueline-app-mien-phi-in-logo-doanh-nghiep/" target="_blank" rel="noreferrer noopener">TẠI ĐÂY</a></em></strong></p>
<h3 id="h-3-loa-bluetooth-mini-recci-l1">3. Loa Bluetooth Mini Recci L1</h3>
<p>Một sự lựa chọn khác tương tự như loa Bluetooth SP005 là Loa Bluetooth Mini Recci L1. Dòng sản phẩm này có thiết kế vô cùng nhỏ gọn, thời trang và giúp khách hàng của doanh nghiệp có thể mang theo mọi lúc mọi nơi. Thời gian sử dụng của loa có thể lên đến 3 giờ khi phát liên tục ở mức âm lượng khoảng 70% với âm thanh Hifi vô cùng sống động. Kết nối Bluetooth của sản phẩm với các thiết bị điện tử hầu như không có độ trễ hay bị nhiễu tần số, giúp người dùng có thể sử dụng cho công việc hoặc tận hưởng âm thanh chất lượng cao.</p>
<div class="wp-block-image"><figure class="aligncenter size-full"><img width="432" height="432" src="https://phukiendienthoaigiare.com/wp-content/uploads/2022/02/Review-Loa-Bluetooth-Mini-Recci-L1-5.jpg" alt="loa Bluetooth cầm tay nhỏ gọn" class="wp-image-17008" srcset="//phukiendienthoaigiare.com/wp-content/uploads/2022/02/Review-Loa-Bluetooth-Mini-Recci-L1-5.jpg 432w, //phukiendienthoaigiare.com/wp-content/uploads/2022/02/Review-Loa-Bluetooth-Mini-Recci-L1-5-300x300.jpg 300w, //phukiendienthoaigiare.com/wp-content/uploads/2022/02/Review-Loa-Bluetooth-Mini-Recci-L1-5-150x150.jpg 150w" sizes="(max-width: 432px) 100vw, 432px" /><figcaption> Loa Bluetooth Mini Recci L1 </figcaption></figure></div>
<p><strong>Thông tin sản phẩm:</strong></p>
<ul><li>Thiết kế nhỏ gọn, thời trang, tiện dụng mang theo.</li><li>Kết nối không dây Bluetooth 4.2 , khoảng cách truyền lên đến 10m.</li><li>Công suất 3w , Âm thanh Hifi sống động.</li><li>Dung lượng pin 400mAh.</li><li>Hỗ trợ thẻ TF và jack 3.5mm.</li><li>Chứng nhận an toàn : CE/RoHS/FCC.</li><li>Phù hợp làm quà tặng Doanh Nghiệp, quà khuyến mãi.</li><li>Bảo Hành chính hãng 6 tháng.</li></ul>
<h3 id="h-4-dế-sạc-khong-day-sạc-nhanh-10w-chuẩn-qi-wct006">4. Đế Sạc Không Dây Sạc Nhanh 10W Chuẩn Qi WCT006</h3>
<p>Đế Sạc Không Dây Sạc Nhanh 10W Chuẩn Qi WCT006 là một sản phẩm khá đặc biệt khi có thể sạc cho tất cả các thiết bị điện thoại có hỗ trợ sạc chuẩn Qi trên thị trường. Khi sử dụng, người dùng sẽ không còn lo lắng về tình trạng sạc quá giờ khiến điện thoại bị chai pin hay nóng thiết bị nhờ vào tính năng tự động ngắt nhiệt cùng khả năng bảo vệ quá nhiệt, quá tải thông minh của sản phẩm. Với thiết kế siêu mỏng, gọn nhẹ, khách hàng của doanh nghiệp có thể mang theo đến bất cứ đâu một cách dễ dàng, vừa phục vụ cho công việc vừa hỗ trợ công tác giải trí của họ.</p>
<div class="wp-block-image"><figure class="aligncenter size-full"><img width="600" height="600" src="https://phukiendienthoaigiare.com/wp-content/uploads/2022/03/Nen-Tang-Qua-Gi-Cho-Khach-Hang1.jpg" alt="Đế sạc nhanh không dây" class="wp-image-18069" srcset="//phukiendienthoaigiare.com/wp-content/uploads/2022/03/Nen-Tang-Qua-Gi-Cho-Khach-Hang1.jpg 600w, //phukiendienthoaigiare.com/wp-content/uploads/2022/03/Nen-Tang-Qua-Gi-Cho-Khach-Hang1-300x300.jpg 300w, //phukiendienthoaigiare.com/wp-content/uploads/2022/03/Nen-Tang-Qua-Gi-Cho-Khach-Hang1-150x150.jpg 150w, //phukiendienthoaigiare.com/wp-content/uploads/2022/03/Nen-Tang-Qua-Gi-Cho-Khach-Hang1-441x441.jpg 441w" sizes="(max-width: 600px) 100vw, 600px" /><figcaption> Đế Sạc Không Dây Sạc Nhanh 10W Chuẩn Qi WCT006 </figcaption></figure></div>
<p><strong>Thông tin sản phẩm:</strong></p>
<ul><li>Đèn LED hiển thị dung lượng pin.</li><li>Sạc không dây 10W/ 7.5W tương thích với tất cả các thiết bị điện thoại đang có trên thị trường.</li><li>Tính năng tự động ngắt sau khi sạc đầy pin cùng khả năng bảo vệ quá nhiệt, quá tải thông minh.</li><li>Thiết kế siêu mỏng, gọn nhẹ dễ dàng mang theo.</li><li>Chứng nhận an toàn: CE.</li><li>Chứng nhận xuất xứ hàng hóa: CO.</li><li>Nhập khẩu và bảo hành 12 tháng theo quy định của nhà sản xuất.</li><li>Phù hợp làm quà tặng Doanh Nghiệp.</li><li>Miễn Phí Tùy Chỉnh Logo Doanh nghiệp theo yêu cầu (MOQ=100pcs).</li></ul>
<h3 id="h-5-pin-sạc-dự-phong-10000-mah-recci-rpb-p09-sạc-nhanh-22-5w-chuẩn-pd-qc3-0">5. Pin Sạc Dự Phòng 10000 mAh Recci RPB-P09 Sạc Nhanh 22.5W Chuẩn PD, QC3.0</h3>
<p>Là người anh em của Đế Sạc Không Dây, Pin sạc dự phòng RPB-P09 sở hữu thiết kế sang trọng, nhỏ gọn và mạnh mẽ với hai màu đen, trắng vô cùng thời trang, phù hợp với phong cách hiện đại của người dùng. Sản phẩm được xem như “phiên bản nâng cấp” của các sản phẩm pin sạc dự phòng hiện nay bởi được tích hợp thêm các tính năng sạc nhanh 22.5W chuẩn PD dành cho điện thoại và máy tính bảng, giúp sạc đầy 50% pin chỉ trong vòng 30 phút với cổng ra (output) Quick Charge 3.0. Ngoài ra, sản phẩm còn bảo vệ cho các thiết bị điện tử không bị nóng khi sạc, tăng tuổi thọ pin nhờ tính năng tự động ngắt khi sạc đầy pin cùng khả năng bảo vệ quá tải, quá nhiệt thông minh.</p>
<div class="wp-block-image"><figure class="aligncenter size-full"><img width="600" height="601" src="https://phukiendienthoaigiare.com/wp-content/uploads/2022/03/Nen-Tang-Qua-Gi-Cho-Khach-Hang2.jpg" alt="pin sạc dự phòng hỗ trợ sạc nhanh" class="wp-image-18070" srcset="//phukiendienthoaigiare.com/wp-content/uploads/2022/03/Nen-Tang-Qua-Gi-Cho-Khach-Hang2.jpg 600w, //phukiendienthoaigiare.com/wp-content/uploads/2022/03/Nen-Tang-Qua-Gi-Cho-Khach-Hang2-300x300.jpg 300w, //phukiendienthoaigiare.com/wp-content/uploads/2022/03/Nen-Tang-Qua-Gi-Cho-Khach-Hang2-150x150.jpg 150w, //phukiendienthoaigiare.com/wp-content/uploads/2022/03/Nen-Tang-Qua-Gi-Cho-Khach-Hang2-441x442.jpg 441w" sizes="(max-width: 600px) 100vw, 600px" /><figcaption> Pin Sạc Dự Phòng 10000 mAh Recci RPB-P09 Sạc Nhanh 22.5W Chuẩn PD, QC3.0 </figcaption></figure></div>
<p>Với Pin sạc dự phòng RPB-P09, khách hàng của doanh nghiệp có thể mang theo khi ra ngoài và sạc đầy pin điện thoại khi cần thiết mà không cần lo sợ ảnh hưởng đến công việc hay nhu cầu giải trí của bản thân.</p>
<p><strong>Thông tin sản phẩm:</strong></p>
<ul><li>Pin sạc dự phòng 10000mAh Recci RPB_P09.</li><li>Sạc Nhanh 22.5W Chuẩn PD , QC3.0 Cho Điện Thoại Và Máy Tính Bảng.</li><li>Thiết kế sang trọng, nhỏ gọn và mạnh mẽ với 2 màu đen, trắng thời trang.</li><li>Đèn LED hiển thị dung lượng pin, mỗi vạch tương ứng 25% dung lượng pin.</li><li>Hỗ trợ 3 cổng ra (output) Quick Charge 3.0 giúp sạc nhanh với công suất tới 22.5 W. sạc đầy 50% pin trong vòng 30 phút, giúp sạc nhanh cho các thiết bị.</li><li>Lõi pin Polymer độ bền cao, dung lượng lớn 10.000 mAh.</li><li>Tính năng tự động ngắt sau khi sạc đầy pin cùng khả năng bảo vệ quá nhiệt, quá tải thông minh.</li><li>Chứng nhận an toàn và phòng chống cháy nổ : CE/FCC/RoHS/MSDS/UN38.3.</li><li>Phù hợp làm quà tặng Doanh Nghiệp,bạn bè, Quà khuyến mãi….</li><li>Nhập khẩu và bảo hành chính hãng 12 tháng.</li></ul>
<p>Trên đây là 5 sản phẩm quà tặng công nghệ các doanh nghiệp có thể lựa chọn làm quà tặng cho khách hàng. Để in/khắc logo thương hiệu, Quý doanh nghiệp vui lòng liên hệ đến hotline của Shop để nhân viên tư vấn và hỗ trợ.</p>
<p>The post <a rel="nofollow" href="https://phukiendienthoaigiare.com/nen-tang-qua-gi-cho-khach-hang/">Nên Tặng Quà Gì Cho Khách Hàng?</a> appeared first on <a rel="nofollow" href="https://phukiendienthoaigiare.com">Phụ kiện điện thoại</a>.</p> | 1.0 | Nên Tặng Quà Gì Cho Khách Hàng? - <p><strong>Những món quà tặng vừa hữu ích, vừa hợp thời sẽ có thể “đánh động” đến trái tim của khách hàng và hỗ trợ chiến lược Marketing vô cùng hiệu quả đấy!</strong></p>
<p>Tặng quà cho khách hàng là một trong những phương thức kinh doanh vô cùng hiệu quả. Doanh nghiệp không chỉ có được mối quan hệ tốt với khách hàng, thúc đẩy doanh thu mà còn có thể Marketing thương hiệu một cách tự nhiên. Song, để có thể đạt được tính PR thương hiệu và được nhiều khách hàng yêu thích, không phải món quà nào doanhh nghiệp cũng có thể “tặng đại” mà cần chú trọng vào một số tiêu chí nhất định.</p>
<h2 id="h-cần-quan-tam-gi-khi-chọn-qua-tặng-cho-khach-hang">Cần quan tâm gì khi chọn quà tặng cho khách hàng?</h2>
<p>Có thể nói, việc tặng quà cho khách hàng đã trở thành một chiến lược kinh doanh không thể thiếu trong hầu hết các doanh nghiệp hiện nay. Tùy thuộc vào ngân sách cũng như sản phẩm/dịch vụ đang cung cấp mà giá trị quà tặng có thể lớn hoặc nhỏ. Tuy nhiên, để chiến lược kinh doanh trở nên hiệu quả, có thể duy trì mối quan hệ trong làm ăn (đối với đối tác, khách hàng là doanh nghiệp) hoặc tri ân khách hàng, thúc đẩy tiêu thụ sản phẩm/sử dụng dịch vụ (người tiêu dùng) và thực hiện chiến lược Marketing hiệu quả thì doanh nghiệp nên lựa chọn quà tặng theo các tiêu chí sau:</p>
<h3 id="h-1-chọn-qua-theo-thời-diểm">1. Chọn quà theo thời điểm</h3>
<p>Những món quà tặng theo thời điểm sẽ vừa thể hiện được sự tinh tế, tâm lý và khả năng “đu trend” của doanh nghiệp vừa có thể giúp khách hàng hòa nhập với xu hướng xã hội hiện nay và có thể sử dụng được ngay tại thời điểm đó. Không nhất thiết là những món quà đắt tiền, doanh nghiệp có thể lựa chọn các sản phẩm quà tặng phù hợp với ngân sách của mình, chẳng hạn như áo mưa (tặng vào mùa mưa), quạt mini, bình giữ nhiệt (tặng vào mùa nóng,…),…</p>
<h3 id="h-2-chọn-qua-co-cong-nang-sử-dụng-cao">2. Chọn quà có công năng sử dụng cao</h3>
<p>Ngoại trừ kinh doanh các sản phẩm trang trí hoặc có liên quan đến trang trí, ở những mặt hàng kinh doanh còn lại, doanh nghiệp nên chọn tặng những vật phẩm có công năng sử dụng cao, tần suất sử dụng nhiều và bất kỳ ai cũng có thể dùng được. Điều này sẽ giúp các doanh nghiệp có thể vừa “phủ sóng” thương hiệu của mình vừa mang đến sự hài lòng cho khách hàng, đạt được hiệu quả trong chiến lược kinh doanh.</p>
<h3 id="h-3-chọn-qua-tặng-tương-xứng-với-gia-trị-sản-phẩm">3. Chọn quà tặng tương xứng với giá trị sản phẩm</h3>
<p>Với những khách hàng mua các sản phẩm có giá trị cao, sự mong đợi về món quà tặng cũng tương xứng giá trị và có ít nhiều liên quan đến sản phẩm doanh nghiệp đang cung cấp. Chẳng hạn, nếu doanh nghiệp kinh doanh các sản phẩm điện máy thì có thể lựa chọn các quà tặng là những phụ kiện kèm theo. Điều này sẽ giúp khách hàng cảm thấy mình “lời” khi mua sản phẩm của doanh nghiệp này. Đặc biệt quan trọng, khi những doanh nghiệp tặng kèm quà khuyến mái trong chiến dịch quảng cáo, khách hàng sẽ có thể vì món quà tặng đó mà sẵn sàng móc “hầu bao” để mua sản phẩm.</p>
<p>Trên đây là 3 tiêu chí đóng vai trò là “kim chỉ nam” để doanh nghiệp có thể lựa chọn những món quà tặng phù hợp. Trong số đó, tiêu chí thứ 2 vô cùng quan trọng và được đại đa số doanh nghiệp áp dụng.</p>
<h2 id="h-top-5-qua-tặng-doanh-nghiệp-hợp-thời-dại-2022">Top 5 quà tặng doanh nghiệp hợp thời đại 2022</h2>
<p>Với thời đại công nghệ 4.0 đang phát triển mạnh mẽ như hiện nay thì các doanh nghiệp nên lựa chọn các quà tặng công nghệ phục vụ cho chiến lược kinh doanh của mình. Những món quà tặng công nghệ vừa có tần suất sử dụng cao vừa phù hợp với hầu hết các đối tượng hiện nay, có thể “phủ sóng” thương hiệu một cách tự nhiên nhất.</p>
<h3 id="h-1-loa-bluetooth-khong-day-sieu-mỏng-bỏ-tui-sp005">1. Loa Bluetooth Không Dây Siêu Mỏng Bỏ Túi SP005</h3>
<p>Loa Bluetooth Không Dây Siêu Mỏng SP005 là một dòng loa di động có thiết kế cầm tay vô cùng nhỏ gọn, tiện dụng. Khi tặng món quà này cho khách hàng, họ không chỉ có thể tận hưởng giai điệu bài hát với âm thanh sống động trong thời gian phát nhạc lên đến 3 giờ mà còn có thể mang theo mọi lúc, mọi nơi.</p>
<div class="wp-block-image"><figure class="aligncenter size-full"><img width="800" height="800" src="https://phukiendienthoaigiare.com/wp-content/uploads/2022/02/Loa-Bluetooth-Khong-Day-Sieu-Mong-Bo-Tui-SP005-6.jpg" alt="In logo lên loa làm quà tặng khách hàng độc đáo" class="wp-image-17017" srcset="//phukiendienthoaigiare.com/wp-content/uploads/2022/02/Loa-Bluetooth-Khong-Day-Sieu-Mong-Bo-Tui-SP005-6.jpg 800w, //phukiendienthoaigiare.com/wp-content/uploads/2022/02/Loa-Bluetooth-Khong-Day-Sieu-Mong-Bo-Tui-SP005-6-300x300.jpg 300w, //phukiendienthoaigiare.com/wp-content/uploads/2022/02/Loa-Bluetooth-Khong-Day-Sieu-Mong-Bo-Tui-SP005-6-150x150.jpg 150w, //phukiendienthoaigiare.com/wp-content/uploads/2022/02/Loa-Bluetooth-Khong-Day-Sieu-Mong-Bo-Tui-SP005-6-768x768.jpg 768w, //phukiendienthoaigiare.com/wp-content/uploads/2022/02/Loa-Bluetooth-Khong-Day-Sieu-Mong-Bo-Tui-SP005-6-441x441.jpg 441w, //phukiendienthoaigiare.com/wp-content/uploads/2022/02/Loa-Bluetooth-Khong-Day-Sieu-Mong-Bo-Tui-SP005-6-600x600.jpg 600w" sizes="(max-width: 800px) 100vw, 800px" /><figcaption> Loa Bluetooth Không Dây Siêu Mỏng Bỏ Túi SP005 </figcaption></figure></div>
<p>Được tích hợp công nghệ kết nối không dây Bluetooth 5.0, chiếc loa Bluetooth SP005 này có thể kết nối nhanh chóng với bất kỳ thiết bị điện tử nào (điện thoại, máy tính bảng, laptop,…) và phát ra những âm thanh to, rõ, trong trẻo trong bán kính 10m. Sản phẩm sở hữu công suất 3W cùng dung lượng pin “khủng 4000 mAh nên có thể hoạt động trong thời gian dài, phù hợp sử dụng trong các chuyến du lịch, team-building cùng đồng nghiệp, bạn bè, người thân,…</p>
<p>Loa Bluetooth SP005 đã đạt được chứng nhận an toàn CE/FCC/RoHS nên luôn đảm bảo an toàn cho người sử dụng. Đặc biệt, nếu các doanh nghiệp đặt mua sản phẩm này tại Shop sẽ được miễn phí in/khắc logo lên bề mặt, phục vụ cho chiến lược Marketing và PR thương hiệu.</p>
<h3 id="h-2-tai-nghe-airpods-1-1-huilian-a6-pro">2. Tai Nghe Airpods 1:1 Huilian A6 Pro</h3>
<p>Tai Nghe Airpods 1:1 Huilian A6 Pro được xem như “quà tặng quốc dân” khi được các doanh nghiệp lựa chọn làm quà tặng công nghệ hiện nay. Không chỉ có công năng sử dụng cao mà chi phí đặt mua chiếc tai nghe nhét tai này lại khá thấp, phù hợp với ngân sách của nhiều doanh nghiệp vừa và nhỏ. Hơn thế nữa, những tính năng được tích hợp bên trong chiếc tai nghe này sẽ có thể khiến cho khách hàng của doanh nghiệp cảm thấy ưng ý.</p>
<div class="wp-block-image"><figure class="aligncenter size-full"><img width="600" height="600" src="https://phukiendienthoaigiare.com/wp-content/uploads/2022/03/Nen-Tang-Qua-Gi-Cho-Khach-Hang.png" alt="Tai Nghe Airpods 1:1 Huilian A6 Pro - Quà tặng giá trị cho khách hàng" class="wp-image-18068" srcset="//phukiendienthoaigiare.com/wp-content/uploads/2022/03/Nen-Tang-Qua-Gi-Cho-Khach-Hang.png 600w, //phukiendienthoaigiare.com/wp-content/uploads/2022/03/Nen-Tang-Qua-Gi-Cho-Khach-Hang-300x300.png 300w, //phukiendienthoaigiare.com/wp-content/uploads/2022/03/Nen-Tang-Qua-Gi-Cho-Khach-Hang-150x150.png 150w, //phukiendienthoaigiare.com/wp-content/uploads/2022/03/Nen-Tang-Qua-Gi-Cho-Khach-Hang-441x441.png 441w" sizes="(max-width: 600px) 100vw, 600px" /><figcaption> Tai Nghe Airpods 1:1 Huilian A6 Pro </figcaption></figure></div>
<p>Tai nghe Bluetooth 5.0 Airpods 1:1 Huilian A6 Pro sở hữu thiết kế nhét tai tiên tiến cùng kiểu dáng thời trang, trẻ trung và hiện đại, phù hợp với xu hướng hiện đại hóa và nhịp sống sôi động hiện nay. Do sản phẩm không có dây như kiểu tai nghe nhét tai thông thường nên người dùng có thể điều khiển thông qua cảm ứng được đặt ở bên trong mỗi chiếc tai nghe. Khi sử dụng sản phẩm này, khách hàng của doanh nghiệp có thể nghe nhạc, giải trí, thực hiện cuộc gọi vô cùng dễ dàng với âm thanh trong trẻo, rõ ràng, chất lượng HD ngay cả khi ở những nơi đông người nhờ công nghệ chống ồn tiên tiến. Tính năng này sẽ phù hợp với những ai thường xuyên ra ngoài làm việc hoặc thực hiện các hoạt động ngoài trời.</p>
<p>Tương tự như loa Bluetooth SP005, tai nghe Airpods 1:1 Huilian A6 Pro sở hữu công nghệ 5.0 tiên tiến, không có độ trễ và khả năng chống nhiễu mạnh. Sản phẩm cũng được làm từ chất liệu cao cấp nên khi thực hiện in/khắc logo lên bề mặt sẽ hiển thị những đường nét rõ ràng, hoàn hảo mà không có bất kỳ trầy xước hay hư hỏng gì.</p>
<p><strong>Thông tin sản phẩm:</strong></p>
<ul><li>Kết nối không dây Bluetooth 5.0 không có độ trễ, chống nhiễu mạnh.</li><li>Đàm thoại 2 tai chất lượng cuộc gọi HD cùng công nghệ chống ồn tiên tiến.</li><li>Mở nắp hiện kết nối hiện popup.</li><li>Tính năng đổi tên & định vị tai trong setting điện thoại.</li><li>Hộp sạc hỗ trợ sạc không dây chuẩn Qi, tiện lợi khi sử dụng.</li><li>Bản lề thép chống gập.</li><li>Phù hợp làm quà tặng Doanh Nghiệp, Bạn Bè, Quà khuyến mãi…</li><li>Tùy Chỉnh Logo Doanh nghiệp theo yêu cầu (MOQ=100pcs).</li></ul>
<p><strong>Mua sản phẩm giá rẻ in logo miễn phí cho Doanh nghiệp: <em><a href="https://phukiendienthoaigiare.com/product/tai-nghe-nhet-tai-in-ear-bluetooth-tws-5-0-jacqueline-app-mien-phi-in-logo-doanh-nghiep/" target="_blank" rel="noreferrer noopener">TẠI ĐÂY</a></em></strong></p>
<h3 id="h-3-loa-bluetooth-mini-recci-l1">3. Loa Bluetooth Mini Recci L1</h3>
<p>Một sự lựa chọn khác tương tự như loa Bluetooth SP005 là Loa Bluetooth Mini Recci L1. Dòng sản phẩm này có thiết kế vô cùng nhỏ gọn, thời trang và giúp khách hàng của doanh nghiệp có thể mang theo mọi lúc mọi nơi. Thời gian sử dụng của loa có thể lên đến 3 giờ khi phát liên tục ở mức âm lượng khoảng 70% với âm thanh Hifi vô cùng sống động. Kết nối Bluetooth của sản phẩm với các thiết bị điện tử hầu như không có độ trễ hay bị nhiễu tần số, giúp người dùng có thể sử dụng cho công việc hoặc tận hưởng âm thanh chất lượng cao.</p>
<div class="wp-block-image"><figure class="aligncenter size-full"><img width="432" height="432" src="https://phukiendienthoaigiare.com/wp-content/uploads/2022/02/Review-Loa-Bluetooth-Mini-Recci-L1-5.jpg" alt="loa Bluetooth cầm tay nhỏ gọn" class="wp-image-17008" srcset="//phukiendienthoaigiare.com/wp-content/uploads/2022/02/Review-Loa-Bluetooth-Mini-Recci-L1-5.jpg 432w, //phukiendienthoaigiare.com/wp-content/uploads/2022/02/Review-Loa-Bluetooth-Mini-Recci-L1-5-300x300.jpg 300w, //phukiendienthoaigiare.com/wp-content/uploads/2022/02/Review-Loa-Bluetooth-Mini-Recci-L1-5-150x150.jpg 150w" sizes="(max-width: 432px) 100vw, 432px" /><figcaption> Loa Bluetooth Mini Recci L1 </figcaption></figure></div>
<p><strong>Thông tin sản phẩm:</strong></p>
<ul><li>Thiết kế nhỏ gọn, thời trang, tiện dụng mang theo.</li><li>Kết nối không dây Bluetooth 4.2 , khoảng cách truyền lên đến 10m.</li><li>Công suất 3w , Âm thanh Hifi sống động.</li><li>Dung lượng pin 400mAh.</li><li>Hỗ trợ thẻ TF và jack 3.5mm.</li><li>Chứng nhận an toàn : CE/RoHS/FCC.</li><li>Phù hợp làm quà tặng Doanh Nghiệp, quà khuyến mãi.</li><li>Bảo Hành chính hãng 6 tháng.</li></ul>
<h3 id="h-4-dế-sạc-khong-day-sạc-nhanh-10w-chuẩn-qi-wct006">4. Đế Sạc Không Dây Sạc Nhanh 10W Chuẩn Qi WCT006</h3>
<p>Đế Sạc Không Dây Sạc Nhanh 10W Chuẩn Qi WCT006 là một sản phẩm khá đặc biệt khi có thể sạc cho tất cả các thiết bị điện thoại có hỗ trợ sạc chuẩn Qi trên thị trường. Khi sử dụng, người dùng sẽ không còn lo lắng về tình trạng sạc quá giờ khiến điện thoại bị chai pin hay nóng thiết bị nhờ vào tính năng tự động ngắt nhiệt cùng khả năng bảo vệ quá nhiệt, quá tải thông minh của sản phẩm. Với thiết kế siêu mỏng, gọn nhẹ, khách hàng của doanh nghiệp có thể mang theo đến bất cứ đâu một cách dễ dàng, vừa phục vụ cho công việc vừa hỗ trợ công tác giải trí của họ.</p>
<div class="wp-block-image"><figure class="aligncenter size-full"><img width="600" height="600" src="https://phukiendienthoaigiare.com/wp-content/uploads/2022/03/Nen-Tang-Qua-Gi-Cho-Khach-Hang1.jpg" alt="Đế sạc nhanh không dây" class="wp-image-18069" srcset="//phukiendienthoaigiare.com/wp-content/uploads/2022/03/Nen-Tang-Qua-Gi-Cho-Khach-Hang1.jpg 600w, //phukiendienthoaigiare.com/wp-content/uploads/2022/03/Nen-Tang-Qua-Gi-Cho-Khach-Hang1-300x300.jpg 300w, //phukiendienthoaigiare.com/wp-content/uploads/2022/03/Nen-Tang-Qua-Gi-Cho-Khach-Hang1-150x150.jpg 150w, //phukiendienthoaigiare.com/wp-content/uploads/2022/03/Nen-Tang-Qua-Gi-Cho-Khach-Hang1-441x441.jpg 441w" sizes="(max-width: 600px) 100vw, 600px" /><figcaption> Đế Sạc Không Dây Sạc Nhanh 10W Chuẩn Qi WCT006 </figcaption></figure></div>
<p><strong>Thông tin sản phẩm:</strong></p>
<ul><li>Đèn LED hiển thị dung lượng pin.</li><li>Sạc không dây 10W/ 7.5W tương thích với tất cả các thiết bị điện thoại đang có trên thị trường.</li><li>Tính năng tự động ngắt sau khi sạc đầy pin cùng khả năng bảo vệ quá nhiệt, quá tải thông minh.</li><li>Thiết kế siêu mỏng, gọn nhẹ dễ dàng mang theo.</li><li>Chứng nhận an toàn: CE.</li><li>Chứng nhận xuất xứ hàng hóa: CO.</li><li>Nhập khẩu và bảo hành 12 tháng theo quy định của nhà sản xuất.</li><li>Phù hợp làm quà tặng Doanh Nghiệp.</li><li>Miễn Phí Tùy Chỉnh Logo Doanh nghiệp theo yêu cầu (MOQ=100pcs).</li></ul>
<h3 id="h-5-pin-sạc-dự-phong-10000-mah-recci-rpb-p09-sạc-nhanh-22-5w-chuẩn-pd-qc3-0">5. Pin Sạc Dự Phòng 10000 mAh Recci RPB-P09 Sạc Nhanh 22.5W Chuẩn PD, QC3.0</h3>
<p>Là người anh em của Đế Sạc Không Dây, Pin sạc dự phòng RPB-P09 sở hữu thiết kế sang trọng, nhỏ gọn và mạnh mẽ với hai màu đen, trắng vô cùng thời trang, phù hợp với phong cách hiện đại của người dùng. Sản phẩm được xem như “phiên bản nâng cấp” của các sản phẩm pin sạc dự phòng hiện nay bởi được tích hợp thêm các tính năng sạc nhanh 22.5W chuẩn PD dành cho điện thoại và máy tính bảng, giúp sạc đầy 50% pin chỉ trong vòng 30 phút với cổng ra (output) Quick Charge 3.0. Ngoài ra, sản phẩm còn bảo vệ cho các thiết bị điện tử không bị nóng khi sạc, tăng tuổi thọ pin nhờ tính năng tự động ngắt khi sạc đầy pin cùng khả năng bảo vệ quá tải, quá nhiệt thông minh.</p>
<div class="wp-block-image"><figure class="aligncenter size-full"><img width="600" height="601" src="https://phukiendienthoaigiare.com/wp-content/uploads/2022/03/Nen-Tang-Qua-Gi-Cho-Khach-Hang2.jpg" alt="pin sạc dự phòng hỗ trợ sạc nhanh" class="wp-image-18070" srcset="//phukiendienthoaigiare.com/wp-content/uploads/2022/03/Nen-Tang-Qua-Gi-Cho-Khach-Hang2.jpg 600w, //phukiendienthoaigiare.com/wp-content/uploads/2022/03/Nen-Tang-Qua-Gi-Cho-Khach-Hang2-300x300.jpg 300w, //phukiendienthoaigiare.com/wp-content/uploads/2022/03/Nen-Tang-Qua-Gi-Cho-Khach-Hang2-150x150.jpg 150w, //phukiendienthoaigiare.com/wp-content/uploads/2022/03/Nen-Tang-Qua-Gi-Cho-Khach-Hang2-441x442.jpg 441w" sizes="(max-width: 600px) 100vw, 600px" /><figcaption> Pin Sạc Dự Phòng 10000 mAh Recci RPB-P09 Sạc Nhanh 22.5W Chuẩn PD, QC3.0 </figcaption></figure></div>
<p>Với Pin sạc dự phòng RPB-P09, khách hàng của doanh nghiệp có thể mang theo khi ra ngoài và sạc đầy pin điện thoại khi cần thiết mà không cần lo sợ ảnh hưởng đến công việc hay nhu cầu giải trí của bản thân.</p>
<p><strong>Thông tin sản phẩm:</strong></p>
<ul><li>Pin sạc dự phòng 10000mAh Recci RPB_P09.</li><li>Sạc Nhanh 22.5W Chuẩn PD , QC3.0 Cho Điện Thoại Và Máy Tính Bảng.</li><li>Thiết kế sang trọng, nhỏ gọn và mạnh mẽ với 2 màu đen, trắng thời trang.</li><li>Đèn LED hiển thị dung lượng pin, mỗi vạch tương ứng 25% dung lượng pin.</li><li>Hỗ trợ 3 cổng ra (output) Quick Charge 3.0 giúp sạc nhanh với công suất tới 22.5 W. sạc đầy 50% pin trong vòng 30 phút, giúp sạc nhanh cho các thiết bị.</li><li>Lõi pin Polymer độ bền cao, dung lượng lớn 10.000 mAh.</li><li>Tính năng tự động ngắt sau khi sạc đầy pin cùng khả năng bảo vệ quá nhiệt, quá tải thông minh.</li><li>Chứng nhận an toàn và phòng chống cháy nổ : CE/FCC/RoHS/MSDS/UN38.3.</li><li>Phù hợp làm quà tặng Doanh Nghiệp,bạn bè, Quà khuyến mãi….</li><li>Nhập khẩu và bảo hành chính hãng 12 tháng.</li></ul>
<p>Trên đây là 5 sản phẩm quà tặng công nghệ các doanh nghiệp có thể lựa chọn làm quà tặng cho khách hàng. Để in/khắc logo thương hiệu, Quý doanh nghiệp vui lòng liên hệ đến hotline của Shop để nhân viên tư vấn và hỗ trợ.</p>
<p>The post <a rel="nofollow" href="https://phukiendienthoaigiare.com/nen-tang-qua-gi-cho-khach-hang/">Nên Tặng Quà Gì Cho Khách Hàng?</a> appeared first on <a rel="nofollow" href="https://phukiendienthoaigiare.com">Phụ kiện điện thoại</a>.</p> | non_infrastructure | nên tặng quà gì cho khách hàng những món quà tặng vừa hữu ích vừa hợp thời sẽ có thể “đánh động” đến trái tim của khách hàng và hỗ trợ chiến lược marketing vô cùng hiệu quả đấy tặng quà cho khách hàng là một trong những phương thức kinh doanh vô cùng hiệu quả doanh nghiệp không chỉ có được mối quan hệ tốt với khách hàng thúc đẩy doanh thu mà còn có thể marketing thương hiệu một cách tự nhiên song để có thể đạt được tính pr thương hiệu và được nhiều khách hàng yêu thích không phải món quà nào doanhh nghiệp cũng có thể “tặng đại” mà cần chú trọng vào một số tiêu chí nhất định cần quan tâm gì khi chọn quà tặng cho khách hàng có thể nói việc tặng quà cho khách hàng đã trở thành một chiến lược kinh doanh không thể thiếu trong hầu hết các doanh nghiệp hiện nay tùy thuộc vào ngân sách cũng như sản phẩm dịch vụ đang cung cấp mà giá trị quà tặng có thể lớn hoặc nhỏ tuy nhiên để chiến lược kinh doanh trở nên hiệu quả có thể duy trì mối quan hệ trong làm ăn đối với đối tác khách hàng là doanh nghiệp hoặc tri ân khách hàng thúc đẩy tiêu thụ sản phẩm sử dụng dịch vụ người tiêu dùng và thực hiện chiến lược marketing hiệu quả thì doanh nghiệp nên lựa chọn quà tặng theo các tiêu chí sau chọn quà theo thời điểm những món quà tặng theo thời điểm sẽ vừa thể hiện được sự tinh tế tâm lý và khả năng “đu trend” của doanh nghiệp vừa có thể giúp khách hàng hòa nhập với xu hướng xã hội hiện nay và có thể sử dụng được ngay tại thời điểm đó không nhất thiết là những món quà đắt tiền doanh nghiệp có thể lựa chọn các sản phẩm quà tặng phù hợp với ngân sách của mình chẳng hạn như áo mưa tặng vào mùa mưa quạt mini bình giữ nhiệt tặng vào mùa nóng … … chọn quà có công năng sử dụng cao ngoại trừ kinh doanh các sản phẩm trang trí hoặc có liên quan đến trang trí ở những mặt hàng kinh doanh còn lại doanh nghiệp nên chọn tặng những vật phẩm có công năng sử dụng cao tần suất sử dụng nhiều và bất kỳ ai cũng có thể dùng được điều này sẽ giúp các doanh nghiệp có thể vừa “phủ sóng” thương hiệu của mình vừa mang đến sự hài lòng cho khách hàng đạt được hiệu quả trong chiến lược kinh doanh chọn quà tặng tương xứng với giá trị sản phẩm với những khách hàng mua các sản phẩm có giá trị cao sự mong đợi về món quà tặng cũng tương xứng giá trị và có ít nhiều liên quan đến sản phẩm doanh nghiệp đang cung cấp chẳng hạn nếu doanh nghiệp kinh doanh các sản phẩm điện máy thì có thể lựa chọn các quà tặng là những phụ kiện kèm theo điều này sẽ giúp khách hàng cảm thấy mình “lời” khi mua sản phẩm của doanh nghiệp này đặc biệt quan trọng khi những doanh nghiệp tặng kèm quà khuyến mái trong chiến dịch quảng cáo khách hàng sẽ có thể vì món quà tặng đó mà sẵn sàng móc “hầu bao” để mua sản phẩm trên đây là tiêu chí đóng vai trò là “kim chỉ nam” để doanh nghiệp có thể lựa chọn những món quà tặng phù hợp trong số đó tiêu chí thứ vô cùng quan trọng và được đại đa số doanh nghiệp áp dụng top quà tặng doanh nghiệp hợp thời đại với thời đại công nghệ đang phát triển mạnh mẽ như hiện nay thì các doanh nghiệp nên lựa chọn các quà tặng công nghệ phục vụ cho chiến lược kinh doanh của mình những món quà tặng công nghệ vừa có tần suất sử dụng cao vừa phù hợp với hầu hết các đối tượng hiện nay có thể “phủ sóng” thương hiệu một cách tự nhiên nhất loa bluetooth không dây siêu mỏng bỏ túi loa bluetooth không dây siêu mỏng là một dòng loa di động có thiết kế cầm tay vô cùng nhỏ gọn tiện dụng khi tặng món quà này cho khách hàng họ không chỉ có thể tận hưởng giai điệu bài hát với âm thanh sống động trong thời gian phát nhạc lên đến giờ mà còn có thể mang theo mọi lúc mọi nơi loa bluetooth không dây siêu mỏng bỏ túi được tích hợp công nghệ kết nối không dây bluetooth chiếc loa bluetooth này có thể kết nối nhanh chóng với bất kỳ thiết bị điện tử nào điện thoại máy tính bảng laptop … và phát ra những âm thanh to rõ trong trẻo trong bán kính sản phẩm sở hữu công suất cùng dung lượng pin “khủng mah nên có thể hoạt động trong thời gian dài phù hợp sử dụng trong các chuyến du lịch team building cùng đồng nghiệp bạn bè người thân … loa bluetooth đã đạt được chứng nhận an toàn ce fcc rohs nên luôn đảm bảo an toàn cho người sử dụng đặc biệt nếu các doanh nghiệp đặt mua sản phẩm này tại shop sẽ được miễn phí in khắc logo lên bề mặt phục vụ cho chiến lược marketing và pr thương hiệu tai nghe airpods huilian pro tai nghe airpods huilian pro được xem như “quà tặng quốc dân” khi được các doanh nghiệp lựa chọn làm quà tặng công nghệ hiện nay không chỉ có công năng sử dụng cao mà chi phí đặt mua chiếc tai nghe nhét tai này lại khá thấp phù hợp với ngân sách của nhiều doanh nghiệp vừa và nhỏ hơn thế nữa những tính năng được tích hợp bên trong chiếc tai nghe này sẽ có thể khiến cho khách hàng của doanh nghiệp cảm thấy ưng ý tai nghe airpods huilian pro tai nghe bluetooth airpods huilian pro sở hữu thiết kế nhét tai tiên tiến cùng kiểu dáng thời trang trẻ trung và hiện đại phù hợp với xu hướng hiện đại hóa và nhịp sống sôi động hiện nay do sản phẩm không có dây như kiểu tai nghe nhét tai thông thường nên người dùng có thể điều khiển thông qua cảm ứng được đặt ở bên trong mỗi chiếc tai nghe khi sử dụng sản phẩm này khách hàng của doanh nghiệp có thể nghe nhạc giải trí thực hiện cuộc gọi vô cùng dễ dàng với âm thanh trong trẻo rõ ràng chất lượng hd ngay cả khi ở những nơi đông người nhờ công nghệ chống ồn tiên tiến tính năng này sẽ phù hợp với những ai thường xuyên ra ngoài làm việc hoặc thực hiện các hoạt động ngoài trời tương tự như loa bluetooth tai nghe airpods huilian pro sở hữu công nghệ tiên tiến không có độ trễ và khả năng chống nhiễu mạnh sản phẩm cũng được làm từ chất liệu cao cấp nên khi thực hiện in khắc logo lên bề mặt sẽ hiển thị những đường nét rõ ràng hoàn hảo mà không có bất kỳ trầy xước hay hư hỏng gì thông tin sản phẩm kết nối không dây bluetooth không có độ trễ chống nhiễu mạnh đàm thoại tai chất lượng cuộc gọi hd cùng công nghệ chống ồn tiên tiến mở nắp hiện kết nối hiện popup tính năng đổi tên định vị tai trong setting điện thoại hộp sạc hỗ trợ sạc không dây chuẩn qi tiện lợi khi sử dụng bản lề thép chống gập phù hợp làm quà tặng doanh nghiệp bạn bè quà khuyến mãi… tùy chỉnh logo doanh nghiệp theo yêu cầu moq mua sản phẩm giá rẻ in logo miễn phí cho doanh nghiệp tại đây loa bluetooth mini recci một sự lựa chọn khác tương tự như loa bluetooth là loa bluetooth mini recci dòng sản phẩm này có thiết kế vô cùng nhỏ gọn thời trang và giúp khách hàng của doanh nghiệp có thể mang theo mọi lúc mọi nơi thời gian sử dụng của loa có thể lên đến giờ khi phát liên tục ở mức âm lượng khoảng với âm thanh hifi vô cùng sống động kết nối bluetooth của sản phẩm với các thiết bị điện tử hầu như không có độ trễ hay bị nhiễu tần số giúp người dùng có thể sử dụng cho công việc hoặc tận hưởng âm thanh chất lượng cao loa bluetooth mini recci thông tin sản phẩm thiết kế nhỏ gọn thời trang tiện dụng mang theo kết nối không dây bluetooth khoảng cách truyền lên đến công suất âm thanh hifi sống động dung lượng pin hỗ trợ thẻ tf và jack chứng nhận an toàn ce rohs fcc phù hợp làm quà tặng doanh nghiệp quà khuyến mãi bảo hành chính hãng tháng đế sạc không dây sạc nhanh chuẩn qi đế sạc không dây sạc nhanh chuẩn qi là một sản phẩm khá đặc biệt khi có thể sạc cho tất cả các thiết bị điện thoại có hỗ trợ sạc chuẩn qi trên thị trường khi sử dụng người dùng sẽ không còn lo lắng về tình trạng sạc quá giờ khiến điện thoại bị chai pin hay nóng thiết bị nhờ vào tính năng tự động ngắt nhiệt cùng khả năng bảo vệ quá nhiệt quá tải thông minh của sản phẩm với thiết kế siêu mỏng gọn nhẹ khách hàng của doanh nghiệp có thể mang theo đến bất cứ đâu một cách dễ dàng vừa phục vụ cho công việc vừa hỗ trợ công tác giải trí của họ đế sạc không dây sạc nhanh chuẩn qi thông tin sản phẩm đèn led hiển thị dung lượng pin sạc không dây tương thích với tất cả các thiết bị điện thoại đang có trên thị trường tính năng tự động ngắt sau khi sạc đầy pin cùng khả năng bảo vệ quá nhiệt quá tải thông minh thiết kế siêu mỏng gọn nhẹ dễ dàng mang theo chứng nhận an toàn ce chứng nhận xuất xứ hàng hóa co nhập khẩu và bảo hành tháng theo quy định của nhà sản xuất phù hợp làm quà tặng doanh nghiệp miễn phí tùy chỉnh logo doanh nghiệp theo yêu cầu moq pin sạc dự phòng mah recci rpb sạc nhanh chuẩn pd là người anh em của đế sạc không dây pin sạc dự phòng rpb sở hữu thiết kế sang trọng nhỏ gọn và mạnh mẽ với hai màu đen trắng vô cùng thời trang phù hợp với phong cách hiện đại của người dùng sản phẩm được xem như “phiên bản nâng cấp” của các sản phẩm pin sạc dự phòng hiện nay bởi được tích hợp thêm các tính năng sạc nhanh chuẩn pd dành cho điện thoại và máy tính bảng giúp sạc đầy pin chỉ trong vòng phút với cổng ra output quick charge ngoài ra sản phẩm còn bảo vệ cho các thiết bị điện tử không bị nóng khi sạc tăng tuổi thọ pin nhờ tính năng tự động ngắt khi sạc đầy pin cùng khả năng bảo vệ quá tải quá nhiệt thông minh pin sạc dự phòng mah recci rpb sạc nhanh chuẩn pd với pin sạc dự phòng rpb khách hàng của doanh nghiệp có thể mang theo khi ra ngoài và sạc đầy pin điện thoại khi cần thiết mà không cần lo sợ ảnh hưởng đến công việc hay nhu cầu giải trí của bản thân thông tin sản phẩm pin sạc dự phòng recci rpb sạc nhanh chuẩn pd cho điện thoại và máy tính bảng thiết kế sang trọng nhỏ gọn và mạnh mẽ với màu đen trắng thời trang đèn led hiển thị dung lượng pin mỗi vạch tương ứng dung lượng pin hỗ trợ cổng ra output quick charge giúp sạc nhanh với công suất tới w sạc đầy pin trong vòng phút giúp sạc nhanh cho các thiết bị lõi pin polymer độ bền cao dung lượng lớn mah tính năng tự động ngắt sau khi sạc đầy pin cùng khả năng bảo vệ quá nhiệt quá tải thông minh chứng nhận an toàn và phòng chống cháy nổ ce fcc rohs msds phù hợp làm quà tặng doanh nghiệp bạn bè quà khuyến mãi… nhập khẩu và bảo hành chính hãng tháng trên đây là sản phẩm quà tặng công nghệ các doanh nghiệp có thể lựa chọn làm quà tặng cho khách hàng để in khắc logo thương hiệu quý doanh nghiệp vui lòng liên hệ đến hotline của shop để nhân viên tư vấn và hỗ trợ the post appeared first on | 0 |
28,901 | 23,595,918,657 | IssuesEvent | 2022-08-23 19:13:43 | carbon-language/carbon-lang | https://api.github.com/repos/carbon-language/carbon-lang | closed | new_proposal.py: could not request reviewer: 'carbon-language/carbon-leads' | infrastructure | I tried to create a new proposal and I got the following error:
```
RUNNING: /usr/bin/gh pr create --draft --label proposal --project Proposals --reviewer carbon-language/carbon-leads --repo carbon-language/carbon-lang --title 'Multidimensional array' --body 'TODO: add summary and links here'
Warning: 1 uncommitted change
could not request reviewer: 'carbon-language/carbon-leads' not found
ERROR: Command failed: /usr/bin/gh pr create --draft --label proposal --project Proposals --reviewer carbon-language/carbon-leads --repo carbon-language/carbon-lang --title 'Multidimensional array' --body 'TODO: add summary and links here'
```
After removing `--reviewer carbon-language/carbon-leads`, PR was successfully created.
My version of `gh` is:
```
$ gh --version
gh version 2.14.2 (2022-07-15)
https://github.com/cli/cli/releases/tag/v2.14.2
``` | 1.0 | new_proposal.py: could not request reviewer: 'carbon-language/carbon-leads' - I tried to create a new proposal and I got the following error:
```
RUNNING: /usr/bin/gh pr create --draft --label proposal --project Proposals --reviewer carbon-language/carbon-leads --repo carbon-language/carbon-lang --title 'Multidimensional array' --body 'TODO: add summary and links here'
Warning: 1 uncommitted change
could not request reviewer: 'carbon-language/carbon-leads' not found
ERROR: Command failed: /usr/bin/gh pr create --draft --label proposal --project Proposals --reviewer carbon-language/carbon-leads --repo carbon-language/carbon-lang --title 'Multidimensional array' --body 'TODO: add summary and links here'
```
After removing `--reviewer carbon-language/carbon-leads`, PR was successfully created.
My version of `gh` is:
```
$ gh --version
gh version 2.14.2 (2022-07-15)
https://github.com/cli/cli/releases/tag/v2.14.2
``` | infrastructure | new proposal py could not request reviewer carbon language carbon leads i tried to create a new proposal and i got the following error running usr bin gh pr create draft label proposal project proposals reviewer carbon language carbon leads repo carbon language carbon lang title multidimensional array body todo add summary and links here warning uncommitted change could not request reviewer carbon language carbon leads not found error command failed usr bin gh pr create draft label proposal project proposals reviewer carbon language carbon leads repo carbon language carbon lang title multidimensional array body todo add summary and links here after removing reviewer carbon language carbon leads pr was successfully created my version of gh is gh version gh version | 1 |
4,918 | 5,331,123,999 | IssuesEvent | 2017-02-15 18:43:16 | morganey-lang/Morganey | https://api.github.com/repos/morganey-lang/Morganey | closed | Error on publishing Morganey 0.0.1 on JCenter | bug infrastructure version: 0.0.1 | Blocks #261
### Reproduction steps
- Switch to branch `release-0.0.1`
- Publish Morganey to JCenter with [bintray-sbt](https://github.com/softprops/bintray-sbt#publishing)
### Observed
An error at the of the publishing:
```
[error] (morganey/*:publish) error uploading to https://api.bintray.com/maven/morganey-lang/maven/maven/me/rexim/morganey_2.11/0.0.1/stdlib_2.11-0.0.1.jar: {"message":"Provided artifact path does not comply with Maven's convention"}
[error] Total time: 33 s, completed Oct 21, 2016 8:30:28 AM
```
### Expected
No errors during the publishing
---
Nothing critical though. stdlib is simply not published. Which doesn't matter because Morganey 0.0.1 cannot load modules from jars anyway.
| 1.0 | Error on publishing Morganey 0.0.1 on JCenter - Blocks #261
### Reproduction steps
- Switch to branch `release-0.0.1`
- Publish Morganey to JCenter with [bintray-sbt](https://github.com/softprops/bintray-sbt#publishing)
### Observed
An error at the of the publishing:
```
[error] (morganey/*:publish) error uploading to https://api.bintray.com/maven/morganey-lang/maven/maven/me/rexim/morganey_2.11/0.0.1/stdlib_2.11-0.0.1.jar: {"message":"Provided artifact path does not comply with Maven's convention"}
[error] Total time: 33 s, completed Oct 21, 2016 8:30:28 AM
```
### Expected
No errors during the publishing
---
Nothing critical though. stdlib is simply not published. Which doesn't matter because Morganey 0.0.1 cannot load modules from jars anyway.
| infrastructure | error on publishing morganey on jcenter blocks reproduction steps switch to branch release publish morganey to jcenter with observed an error at the of the publishing morganey publish error uploading to message provided artifact path does not comply with maven s convention total time s completed oct am expected no errors during the publishing nothing critical though stdlib is simply not published which doesn t matter because morganey cannot load modules from jars anyway | 1 |
50,887 | 7,642,818,516 | IssuesEvent | 2018-05-08 10:30:41 | Microsoft/vsts-tasks | https://api.github.com/repos/Microsoft/vsts-tasks | closed | No documentation for Azure Powershell task | Area: Documentation Area: Release | https://www.visualstudio.com/en-gb/docs/build/steps/deploy/azure-powershell
When I click on the referenced GitHub page I see just source code, no documentation, not even a README file. How is this helping the average user? | 1.0 | No documentation for Azure Powershell task - https://www.visualstudio.com/en-gb/docs/build/steps/deploy/azure-powershell
When I click on the referenced GitHub page I see just source code, no documentation, not even a README file. How is this helping the average user? | non_infrastructure | no documentation for azure powershell task when i click on the referenced github page i see just source code no documentation not even a readme file how is this helping the average user | 0 |
774,937 | 27,215,006,769 | IssuesEvent | 2023-02-20 20:37:15 | ascheid/itsg33-pbmm-issue-gen | https://api.github.com/repos/ascheid/itsg33-pbmm-issue-gen | closed | PL-1 SECURITY PLANNING POLICY AND PROCEDURES | Priority: P1 | (A) The organization develops, documents, and disseminates to [Assignment: organization-defined personnel or roles]:
(a) A security planning policy that addresses purpose, scope, roles, responsibilities, management commitment, coordination among organizational entities, and compliance; and
(b) Procedures to facilitate the implementation of the security planning policy and associated security planning controls.
(B) The organization reviews and updates the current:
(a) Security planning policy [Assignment: organization-defined frequency]; and
(b) Security planning procedures [Assignment: organization-defined frequency]. | 1.0 | PL-1 SECURITY PLANNING POLICY AND PROCEDURES - (A) The organization develops, documents, and disseminates to [Assignment: organization-defined personnel or roles]:
(a) A security planning policy that addresses purpose, scope, roles, responsibilities, management commitment, coordination among organizational entities, and compliance; and
(b) Procedures to facilitate the implementation of the security planning policy and associated security planning controls.
(B) The organization reviews and updates the current:
(a) Security planning policy [Assignment: organization-defined frequency]; and
(b) Security planning procedures [Assignment: organization-defined frequency]. | non_infrastructure | pl security planning policy and procedures a the organization develops documents and disseminates to a a security planning policy that addresses purpose scope roles responsibilities management commitment coordination among organizational entities and compliance and b procedures to facilitate the implementation of the security planning policy and associated security planning controls b the organization reviews and updates the current a security planning policy and b security planning procedures | 0 |
25,560 | 18,852,422,874 | IssuesEvent | 2021-11-11 23:00:30 | dotnet/roslyn | https://api.github.com/repos/dotnet/roslyn | opened | [Flaky Test] Roslyn.VisualStudio.IntegrationTests.VisualBasic.BasicNavigationBar.CodeSpit | Area-Infrastructure Integration-Test Infraswat | Example run: https://runfo.azurewebsites.net/view/build/?number=1459935
Example message:
> Assert.Contains() Failure Not found: Public Sub New() End Sub In value: Class C Public WithEvents Domain As AppDomain Public Sub Goo() End Sub End Class Structure S Public Property A As Integer Public Property B As Integer End Structure
| 1.0 | [Flaky Test] Roslyn.VisualStudio.IntegrationTests.VisualBasic.BasicNavigationBar.CodeSpit - Example run: https://runfo.azurewebsites.net/view/build/?number=1459935
Example message:
> Assert.Contains() Failure Not found: Public Sub New() End Sub In value: Class C Public WithEvents Domain As AppDomain Public Sub Goo() End Sub End Class Structure S Public Property A As Integer Public Property B As Integer End Structure
| infrastructure | roslyn visualstudio integrationtests visualbasic basicnavigationbar codespit example run example message assert contains failure not found public sub new end sub in value class c public withevents domain as appdomain public sub goo end sub end class structure s public property a as integer public property b as integer end structure | 1 |
89,599 | 18,009,140,692 | IssuesEvent | 2021-09-16 06:15:37 | KuznetsovR/Dnevnik | https://api.github.com/repos/KuznetsovR/Dnevnik | opened | Improve test's coverage more then 50% | codestyle | Our coverage utility should to show coverage value more then 50% | 1.0 | Improve test's coverage more then 50% - Our coverage utility should to show coverage value more then 50% | non_infrastructure | improve test s coverage more then our coverage utility should to show coverage value more then | 0 |
340,109 | 30,493,028,905 | IssuesEvent | 2023-07-18 08:59:12 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | roachtest: declarative_schema_changer/job-compatibility-mixed-version-V222-V231 failed | C-test-failure O-robot O-roachtest branch-master release-blocker T-sql-foundations | roachtest.declarative_schema_changer/job-compatibility-mixed-version-V222-V231 [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10950435?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10950435?buildTab=artifacts#/declarative_schema_changer/job-compatibility-mixed-version-V222-V231) on master @ [7675ca4998134028f0623e04737b5cb69fcc33a9](https://github.com/cockroachdb/cockroach/commits/7675ca4998134028f0623e04737b5cb69fcc33a9):
```
(versionupgrade.go:260).func1: ~ COCKROACH_CONNECT_TIMEOUT=1200 ./v23.1.5/cockroach sql --url 'postgres://root@localhost:26257?sslmode=disable' -e "CREATE SCHEDULE IF NOT EXISTS test_only_backup FOR BACKUP INTO 'gs://cockroach-backup-testing-private/roachprod-scheduled-backups/teamcity-10950435-1689659335-14-n4cpu4/1689659964082069869?AUTH=implicit' RECURRING '*/15 * * * *' FULL BACKUP '@hourly' WITH SCHEDULE OPTIONS first_run = 'now'"
ERROR: unexpected error occurred when checking for existing backups in gs://cockroach-backup-testing-private/roachprod-scheduled-backups/teamcity-10950435-1689659335-14-n4cpu4/1689659964082069869?AUTH=implicit: unable to list files in gcs bucket: googleapi: Error 403: 21965078311-compute@developer.gserviceaccount.com does not have storage.objects.list access to the Google Cloud Storage bucket. Permission 'storage.objects.list' denied on resource (or it may not exist).
SQLSTATE: 58030
Failed running "sql": COMMAND_PROBLEM: exit status 1
test artifacts and logs in: /artifacts/declarative_schema_changer/job-compatibility-mixed-version-V222-V231/run_1
```
<p>Parameters: <code>ROACHTEST_arch=amd64</code>
, <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/sql-foundations
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*declarative_schema_changer/job-compatibility-mixed-version-V222-V231.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-29794 | 2.0 | roachtest: declarative_schema_changer/job-compatibility-mixed-version-V222-V231 failed - roachtest.declarative_schema_changer/job-compatibility-mixed-version-V222-V231 [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10950435?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10950435?buildTab=artifacts#/declarative_schema_changer/job-compatibility-mixed-version-V222-V231) on master @ [7675ca4998134028f0623e04737b5cb69fcc33a9](https://github.com/cockroachdb/cockroach/commits/7675ca4998134028f0623e04737b5cb69fcc33a9):
```
(versionupgrade.go:260).func1: ~ COCKROACH_CONNECT_TIMEOUT=1200 ./v23.1.5/cockroach sql --url 'postgres://root@localhost:26257?sslmode=disable' -e "CREATE SCHEDULE IF NOT EXISTS test_only_backup FOR BACKUP INTO 'gs://cockroach-backup-testing-private/roachprod-scheduled-backups/teamcity-10950435-1689659335-14-n4cpu4/1689659964082069869?AUTH=implicit' RECURRING '*/15 * * * *' FULL BACKUP '@hourly' WITH SCHEDULE OPTIONS first_run = 'now'"
ERROR: unexpected error occurred when checking for existing backups in gs://cockroach-backup-testing-private/roachprod-scheduled-backups/teamcity-10950435-1689659335-14-n4cpu4/1689659964082069869?AUTH=implicit: unable to list files in gcs bucket: googleapi: Error 403: 21965078311-compute@developer.gserviceaccount.com does not have storage.objects.list access to the Google Cloud Storage bucket. Permission 'storage.objects.list' denied on resource (or it may not exist).
SQLSTATE: 58030
Failed running "sql": COMMAND_PROBLEM: exit status 1
test artifacts and logs in: /artifacts/declarative_schema_changer/job-compatibility-mixed-version-V222-V231/run_1
```
<p>Parameters: <code>ROACHTEST_arch=amd64</code>
, <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/sql-foundations
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*declarative_schema_changer/job-compatibility-mixed-version-V222-V231.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-29794 | non_infrastructure | roachtest declarative schema changer job compatibility mixed version failed roachtest declarative schema changer job compatibility mixed version with on master versionupgrade go cockroach connect timeout cockroach sql url postgres root localhost sslmode disable e create schedule if not exists test only backup for backup into gs cockroach backup testing private roachprod scheduled backups teamcity auth implicit recurring full backup hourly with schedule options first run now error unexpected error occurred when checking for existing backups in gs cockroach backup testing private roachprod scheduled backups teamcity auth implicit unable to list files in gcs bucket googleapi error compute developer gserviceaccount com does not have storage objects list access to the google cloud storage bucket permission storage objects list denied on resource or it may not exist sqlstate failed running sql command problem exit status test artifacts and logs in artifacts declarative schema changer job compatibility mixed version run parameters roachtest arch roachtest cloud gce roachtest cpu roachtest encrypted false roachtest ssd help see see cc cockroachdb sql foundations jira issue crdb | 0 |
409,640 | 27,746,387,473 | IssuesEvent | 2023-03-15 17:19:24 | stuart-lab/signac | https://api.github.com/repos/stuart-lab/signac | opened | Problem with "Counting fraction of reads in peaks" `FRiP()` in section "Working with datasets that were not quantified using CellRanger" | documentation | <!-- A clear description of what content at the Signac website or in the Signac function man pages is an issue. -->
## Compare between 2 formulas:
#### This is the code (in the Signac document) for calculate fraction of reads in peaks when output data is FROM Cellranger:
```
pbmc$pct_reads_in_peaks <- pbmc$peak_region_fragments / pbmc$passed_filters * 100
```
#### From original documentation of 10X: "peak_region_fragments" means number of fragments overlapping peaks
---
#### This is the code (in the document) for calculate fraction of reads in peaks when output data is NOT FROM Cellranger:
```
# not run
total_fragments <- [CountFragments](https://stuartlab.org/signac/reference/CountFragments.html)("'../vignette_data/atac_v1_pbmc_10k_fragments.tsv.gz'")
pbmc$fragments <- total_fragments[[colnames](https://rdrr.io/pkg/BiocGenerics/man/row_colnames.html)(pbmc), "frequency_count"]
pbmc <- [FRiP](https://stuartlab.org/signac/reference/FRiP.html)(
object = pbmc,
assay = 'peaks',
total.fragments = 'fragments'
)
```
#### I checked the source code of `FRiP`, it will take [sum count of a cell (in matrix)] / pbmc$fragments * 100.
----
### The data from pbmc$passed_filters (from cellranger meta) and pbmc$fragments (calculated) are similar, but the pbmc$peak_region_fragments and [sum count of a cell (in matrix)] are different. And that leads to the "fraction of reads in peaks" are now different between 2 ways of calculation. Can you explain it? | 1.0 | Problem with "Counting fraction of reads in peaks" `FRiP()` in section "Working with datasets that were not quantified using CellRanger" - <!-- A clear description of what content at the Signac website or in the Signac function man pages is an issue. -->
## Compare between 2 formulas:
#### This is the code (in the Signac document) for calculate fraction of reads in peaks when output data is FROM Cellranger:
```
pbmc$pct_reads_in_peaks <- pbmc$peak_region_fragments / pbmc$passed_filters * 100
```
#### From original documentation of 10X: "peak_region_fragments" means number of fragments overlapping peaks
---
#### This is the code (in the document) for calculate fraction of reads in peaks when output data is NOT FROM Cellranger:
```
# not run
total_fragments <- [CountFragments](https://stuartlab.org/signac/reference/CountFragments.html)("'../vignette_data/atac_v1_pbmc_10k_fragments.tsv.gz'")
pbmc$fragments <- total_fragments[[colnames](https://rdrr.io/pkg/BiocGenerics/man/row_colnames.html)(pbmc), "frequency_count"]
pbmc <- [FRiP](https://stuartlab.org/signac/reference/FRiP.html)(
object = pbmc,
assay = 'peaks',
total.fragments = 'fragments'
)
```
#### I checked the source code of `FRiP`, it will take [sum count of a cell (in matrix)] / pbmc$fragments * 100.
----
### The data from pbmc$passed_filters (from cellranger meta) and pbmc$fragments (calculated) are similar, but the pbmc$peak_region_fragments and [sum count of a cell (in matrix)] are different. And that leads to the "fraction of reads in peaks" are now different between 2 ways of calculation. Can you explain it? | non_infrastructure | problem with counting fraction of reads in peaks frip in section working with datasets that were not quantified using cellranger compare between formulas this is the code in the signac document for calculate fraction of reads in peaks when output data is from cellranger pbmc pct reads in peaks pbmc peak region fragments pbmc passed filters from original documentation of peak region fragments means number of fragments overlapping peaks this is the code in the document for calculate fraction of reads in peaks when output data is not from cellranger not run total fragments pbmc fragments total fragments frequency count pbmc object pbmc assay peaks total fragments fragments i checked the source code of frip it will take pbmc fragments the data from pbmc passed filters from cellranger meta and pbmc fragments calculated are similar but the pbmc peak region fragments and are different and that leads to the fraction of reads in peaks are now different between ways of calculation can you explain it | 0 |
640,870 | 20,810,937,603 | IssuesEvent | 2022-03-18 02:31:25 | magma/magma | https://api.github.com/repos/magma/magma | closed | Need to fix probable memory leaks | type: bug wontfix priority: high product: 5g sa |
1) Some of the AMF GRPC request/response and internal messages are not removed in the common function "**itti_free_msg_content**". Please incorporate the missing relevant messages here and call this function from the respective functional code
2) **amf_free_ue_context** this function frees ue context data structures observed that this function is not called from many of the scenarios such as:
i) ue initiated context release command
ii) some of the procedure abort/invalid scenarios
3) For registration procedures, code is allocating heap for the entire struct but while calling delete_procedures, it is removing only the part of the struct
For e.g:
nas5g_amf_auth_proc_t* auth_proc = new (nas5g_amf_auth_proc_t)();
only auth_proc->amf_com_proc this part of the struct is cleaned up
| 1.0 | Need to fix probable memory leaks -
1) Some of the AMF GRPC request/response and internal messages are not removed in the common function "**itti_free_msg_content**". Please incorporate the missing relevant messages here and call this function from the respective functional code
2) **amf_free_ue_context** this function frees ue context data structures observed that this function is not called from many of the scenarios such as:
i) ue initiated context release command
ii) some of the procedure abort/invalid scenarios
3) For registration procedures, code is allocating heap for the entire struct but while calling delete_procedures, it is removing only the part of the struct
For e.g:
nas5g_amf_auth_proc_t* auth_proc = new (nas5g_amf_auth_proc_t)();
only auth_proc->amf_com_proc this part of the struct is cleaned up
| non_infrastructure | need to fix probable memory leaks some of the amf grpc request response and internal messages are not removed in the common function itti free msg content please incorporate the missing relevant messages here and call this function from the respective functional code amf free ue context this function frees ue context data structures observed that this function is not called from many of the scenarios such as i ue initiated context release command ii some of the procedure abort invalid scenarios for registration procedures code is allocating heap for the entire struct but while calling delete procedures it is removing only the part of the struct for e g amf auth proc t auth proc new amf auth proc t only auth proc amf com proc this part of the struct is cleaned up | 0 |
8,227 | 7,298,841,250 | IssuesEvent | 2018-02-26 18:12:04 | dotnet/roslyn-analyzers | https://api.github.com/repos/dotnet/roslyn-analyzers | closed | Review unused cancellation tokens in non-shipping projects | 4 - In Review Area-Infrastructure | We have a 70+ instances of unused cancellation tokens. We need to review these and make sure they're being used. Reenabled CA1801 in the ruleset to view all. | 1.0 | Review unused cancellation tokens in non-shipping projects - We have a 70+ instances of unused cancellation tokens. We need to review these and make sure they're being used. Reenabled CA1801 in the ruleset to view all. | infrastructure | review unused cancellation tokens in non shipping projects we have a instances of unused cancellation tokens we need to review these and make sure they re being used reenabled in the ruleset to view all | 1 |
17,800 | 12,577,646,644 | IssuesEvent | 2020-06-09 09:52:47 | QubesOS/qubes-issues | https://api.github.com/repos/QubesOS/qubes-issues | closed | Error when locally serving Qubes website on Fedora 31 | C: infrastructure C: website P: default T: bug | **Qubes OS version**
N/A
(This is a bug that affects Qubes users who are trying to serve the Qubes website locally, especially those who work on the website, like myself. It is not a bug internal to Qubes OS itself.)
**Affected component(s) or functionality**
Locally serving the Qubes website on Fedora 31 template
**Brief summary**
With my Fedora 30 template, I was able to serve the website locally without a problem. After upgrading to Fedora 31, I encounter an error when trying to serve the website locally, and I can't find any solutions by searching the web for the error message. This hinders my and all contributors' effectiveness in working on the website, documentation, and posts, since I cannot preview the site before pushing live changes.
**To Reproduce**
Steps to reproduce the behavior:
1. Upgrade template from Fedora 30 to 31.
2. Enter Qubes website repo directory.
3. `sudo systemctl start docker && sudo make`
**Expected behavior**
On Fedora 30:
```
$ sudo systemctl start docker && sudo make
mv Gemfile _Gemfile && docker-compose up && mv _Gemfile Gemfile
Pulling jekyll (jekyll/jekyll:pages)...
Trying to pull repository docker.io/jekyll/jekyll ...
sha256:f7a1edb2a9133c690d90cbb97583381b348a56be141e75090dbdcdbc3cc0fd44: Pulling from docker.io/jekyll/jekyll
cbdbe7a5bc2a: Pull complete
aa8ae8202b42: Pull complete
b21786fe7c0d: Pull complete
2e8deee27474: Pull complete
b84a6ba69f44: Pull complete
a4e8966314b7: Pull complete
Digest: sha256:f7a1edb2a9133c690d90cbb97583381b348a56be141e75090dbdcdbc3cc0fd44
Status: Downloaded newer image for docker.io/jekyll/jekyll:pages
Creating qubesosgithubio_jekyll_1 ... done
Attaching to qubesosgithubio_jekyll_1
jekyll_1 | ruby 2.7.1p83 (2020-03-31 revision a0c7c23c9c) [x86_64-linux-musl]
jekyll_1 | Configuration file: /srv/jekyll/_config.yml
jekyll_1 | Source: /srv/jekyll
jekyll_1 | Destination: /srv/jekyll/_site
jekyll_1 | Incremental build: disabled. Enable with --incremental
jekyll_1 | Generating...
jekyll_1 | done in 13.768 seconds.
jekyll_1 | Auto-regeneration: enabled for '/srv/jekyll'
jekyll_1 | Server address: http://0.0.0.0:4000
jekyll_1 | Server running... press ctrl-c to stop.
```
**Actual behavior**
After updating to Fedora 31:
```
$ sudo systemctl start docker && sudo make
mv Gemfile _Gemfile && docker-compose up && mv _Gemfile Gemfile
Pulling jekyll (jekyll/jekyll:pages)...
pages: Pulling from jekyll/jekyll
cbdbe7a5bc2a: Pull complete
aa8ae8202b42: Pull complete
b21786fe7c0d: Pull complete
2e8deee27474: Pull complete
b84a6ba69f44: Pull complete
a4e8966314b7: Pull complete
Digest: sha256:f7a1edb2a9133c690d90cbb97583381b348a56be141e75090dbdcdbc3cc0fd44
Status: Downloaded newer image for jekyll/jekyll:pages
Creating qubesosgithubio_jekyll_1 ... error
ERROR: for qubesosgithubio_jekyll_1 Cannot start service jekyll: OCI runtime create failed: this version of runc doesn't work on cgroups v2: unknown
ERROR: for jekyll Cannot start service jekyll: OCI runtime create failed: this version of runc doesn't work on cgroups v2: unknown
ERROR: Encountered errors while bringing up the project.
make: *** [Makefile:2: all] Error 1
```
**Solutions you've tried**
I've tried searching the web for `Cannot start service jekyll: OCI runtime create failed: this version of runc doesn't work on cgroups v2: unknown` (and parts and variations of this). None of the results have led to a solution. Some of them seem to suggest that it might be an upstream bug, but I'm not sure.
**Relevant [documentation](https://www.qubes-os.org/doc/) you've consulted**
https://github.com/QubesOS/qubesos.github.io#instructions
**Related, [non-duplicate](https://www.qubes-os.org/doc/reporting-bugs/#new-issues-should-not-be-duplicates-of-existing-issues) issues**
None found.
| 1.0 | Error when locally serving Qubes website on Fedora 31 - **Qubes OS version**
N/A
(This is a bug that affects Qubes users who are trying to serve the Qubes website locally, especially those who work on the website, like myself. It is not a bug internal to Qubes OS itself.)
**Affected component(s) or functionality**
Locally serving the Qubes website on Fedora 31 template
**Brief summary**
With my Fedora 30 template, I was able to serve the website locally without a problem. After upgrading to Fedora 31, I encounter an error when trying to serve the website locally, and I can't find any solutions by searching the web for the error message. This hinders my and all contributors' effectiveness in working on the website, documentation, and posts, since I cannot preview the site before pushing live changes.
**To Reproduce**
Steps to reproduce the behavior:
1. Upgrade template from Fedora 30 to 31.
2. Enter Qubes website repo directory.
3. `sudo systemctl start docker && sudo make`
**Expected behavior**
On Fedora 30:
```
$ sudo systemctl start docker && sudo make
mv Gemfile _Gemfile && docker-compose up && mv _Gemfile Gemfile
Pulling jekyll (jekyll/jekyll:pages)...
Trying to pull repository docker.io/jekyll/jekyll ...
sha256:f7a1edb2a9133c690d90cbb97583381b348a56be141e75090dbdcdbc3cc0fd44: Pulling from docker.io/jekyll/jekyll
cbdbe7a5bc2a: Pull complete
aa8ae8202b42: Pull complete
b21786fe7c0d: Pull complete
2e8deee27474: Pull complete
b84a6ba69f44: Pull complete
a4e8966314b7: Pull complete
Digest: sha256:f7a1edb2a9133c690d90cbb97583381b348a56be141e75090dbdcdbc3cc0fd44
Status: Downloaded newer image for docker.io/jekyll/jekyll:pages
Creating qubesosgithubio_jekyll_1 ... done
Attaching to qubesosgithubio_jekyll_1
jekyll_1 | ruby 2.7.1p83 (2020-03-31 revision a0c7c23c9c) [x86_64-linux-musl]
jekyll_1 | Configuration file: /srv/jekyll/_config.yml
jekyll_1 | Source: /srv/jekyll
jekyll_1 | Destination: /srv/jekyll/_site
jekyll_1 | Incremental build: disabled. Enable with --incremental
jekyll_1 | Generating...
jekyll_1 | done in 13.768 seconds.
jekyll_1 | Auto-regeneration: enabled for '/srv/jekyll'
jekyll_1 | Server address: http://0.0.0.0:4000
jekyll_1 | Server running... press ctrl-c to stop.
```
**Actual behavior**
After updating to Fedora 31:
```
$ sudo systemctl start docker && sudo make
mv Gemfile _Gemfile && docker-compose up && mv _Gemfile Gemfile
Pulling jekyll (jekyll/jekyll:pages)...
pages: Pulling from jekyll/jekyll
cbdbe7a5bc2a: Pull complete
aa8ae8202b42: Pull complete
b21786fe7c0d: Pull complete
2e8deee27474: Pull complete
b84a6ba69f44: Pull complete
a4e8966314b7: Pull complete
Digest: sha256:f7a1edb2a9133c690d90cbb97583381b348a56be141e75090dbdcdbc3cc0fd44
Status: Downloaded newer image for jekyll/jekyll:pages
Creating qubesosgithubio_jekyll_1 ... error
ERROR: for qubesosgithubio_jekyll_1 Cannot start service jekyll: OCI runtime create failed: this version of runc doesn't work on cgroups v2: unknown
ERROR: for jekyll Cannot start service jekyll: OCI runtime create failed: this version of runc doesn't work on cgroups v2: unknown
ERROR: Encountered errors while bringing up the project.
make: *** [Makefile:2: all] Error 1
```
**Solutions you've tried**
I've tried searching the web for `Cannot start service jekyll: OCI runtime create failed: this version of runc doesn't work on cgroups v2: unknown` (and parts and variations of this). None of the results have led to a solution. Some of them seem to suggest that it might be an upstream bug, but I'm not sure.
**Relevant [documentation](https://www.qubes-os.org/doc/) you've consulted**
https://github.com/QubesOS/qubesos.github.io#instructions
**Related, [non-duplicate](https://www.qubes-os.org/doc/reporting-bugs/#new-issues-should-not-be-duplicates-of-existing-issues) issues**
None found.
| infrastructure | error when locally serving qubes website on fedora qubes os version n a this is a bug that affects qubes users who are trying to serve the qubes website locally especially those who work on the website like myself it is not a bug internal to qubes os itself affected component s or functionality locally serving the qubes website on fedora template brief summary with my fedora template i was able to serve the website locally without a problem after upgrading to fedora i encounter an error when trying to serve the website locally and i can t find any solutions by searching the web for the error message this hinders my and all contributors effectiveness in working on the website documentation and posts since i cannot preview the site before pushing live changes to reproduce steps to reproduce the behavior upgrade template from fedora to enter qubes website repo directory sudo systemctl start docker sudo make expected behavior on fedora sudo systemctl start docker sudo make mv gemfile gemfile docker compose up mv gemfile gemfile pulling jekyll jekyll jekyll pages trying to pull repository docker io jekyll jekyll pulling from docker io jekyll jekyll pull complete pull complete pull complete pull complete pull complete pull complete digest status downloaded newer image for docker io jekyll jekyll pages creating qubesosgithubio jekyll done attaching to qubesosgithubio jekyll jekyll ruby revision jekyll configuration file srv jekyll config yml jekyll source srv jekyll jekyll destination srv jekyll site jekyll incremental build disabled enable with incremental jekyll generating jekyll done in seconds jekyll auto regeneration enabled for srv jekyll jekyll server address jekyll server running press ctrl c to stop actual behavior after updating to fedora sudo systemctl start docker sudo make mv gemfile gemfile docker compose up mv gemfile gemfile pulling jekyll jekyll jekyll pages pages pulling from jekyll jekyll pull complete pull complete pull complete pull complete pull complete pull complete digest status downloaded newer image for jekyll jekyll pages creating qubesosgithubio jekyll error error for qubesosgithubio jekyll cannot start service jekyll oci runtime create failed this version of runc doesn t work on cgroups unknown error for jekyll cannot start service jekyll oci runtime create failed this version of runc doesn t work on cgroups unknown error encountered errors while bringing up the project make error solutions you ve tried i ve tried searching the web for cannot start service jekyll oci runtime create failed this version of runc doesn t work on cgroups unknown and parts and variations of this none of the results have led to a solution some of them seem to suggest that it might be an upstream bug but i m not sure relevant you ve consulted related issues none found | 1 |
133,543 | 18,298,939,565 | IssuesEvent | 2021-10-05 23:47:37 | bsbtd/Teste | https://api.github.com/repos/bsbtd/Teste | opened | CVE-2020-7753 (High) detected in trim-0.0.1.tgz | security vulnerability | ## CVE-2020-7753 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>trim-0.0.1.tgz</b></p></summary>
<p>Trim string whitespace</p>
<p>Library home page: <a href="https://registry.npmjs.org/trim/-/trim-0.0.1.tgz">https://registry.npmjs.org/trim/-/trim-0.0.1.tgz</a></p>
<p>Path to dependency file: Teste/bootstrap/package.json</p>
<p>Path to vulnerable library: .1.tgz,Teste/bootstrap/node_modules/trim/package.json,Teste/liferay-portal/modules/node_modules/trim/package.json,Teste/pro-table/node_modules/trim/package.json</p>
<p>
Dependency Hierarchy:
- :x: **trim-0.0.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/bsbtd/Teste/commit/64dde89c50c07496423c4d4a865f2e16b92399ad">64dde89c50c07496423c4d4a865f2e16b92399ad</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
All versions of package trim are vulnerable to Regular Expression Denial of Service (ReDoS) via trim().
<p>Publish Date: 2020-10-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7753>CVE-2020-7753</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/component/trim/pull/8">https://github.com/component/trim/pull/8</a></p>
<p>Release Date: 2020-10-27</p>
<p>Fix Resolution: trim - 0.0.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-7753 (High) detected in trim-0.0.1.tgz - ## CVE-2020-7753 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>trim-0.0.1.tgz</b></p></summary>
<p>Trim string whitespace</p>
<p>Library home page: <a href="https://registry.npmjs.org/trim/-/trim-0.0.1.tgz">https://registry.npmjs.org/trim/-/trim-0.0.1.tgz</a></p>
<p>Path to dependency file: Teste/bootstrap/package.json</p>
<p>Path to vulnerable library: .1.tgz,Teste/bootstrap/node_modules/trim/package.json,Teste/liferay-portal/modules/node_modules/trim/package.json,Teste/pro-table/node_modules/trim/package.json</p>
<p>
Dependency Hierarchy:
- :x: **trim-0.0.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/bsbtd/Teste/commit/64dde89c50c07496423c4d4a865f2e16b92399ad">64dde89c50c07496423c4d4a865f2e16b92399ad</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
All versions of package trim are vulnerable to Regular Expression Denial of Service (ReDoS) via trim().
<p>Publish Date: 2020-10-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7753>CVE-2020-7753</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/component/trim/pull/8">https://github.com/component/trim/pull/8</a></p>
<p>Release Date: 2020-10-27</p>
<p>Fix Resolution: trim - 0.0.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | cve high detected in trim tgz cve high severity vulnerability vulnerable library trim tgz trim string whitespace library home page a href path to dependency file teste bootstrap package json path to vulnerable library tgz teste bootstrap node modules trim package json teste liferay portal modules node modules trim package json teste pro table node modules trim package json dependency hierarchy x trim tgz vulnerable library found in head commit a href vulnerability details all versions of package trim are vulnerable to regular expression denial of service redos via trim publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution trim step up your open source security game with whitesource | 0 |
320,842 | 23,828,511,683 | IssuesEvent | 2022-09-05 17:12:14 | n1ru4l/envelop | https://api.github.com/repos/n1ru4l/envelop | closed | useAsyncSchema does not await the schema promise | documentation kind/question | **Describe the bug**
The `useAsyncSchema` accepts a promise that resolves to a GraphQL schema as argument. It chains a `.then()` on this promise (inside the call to `setSchema` happens), but since the envelop initialisation is synchronous, it might happen that the GraphQL server is handling the first request before the schema promise has resolved. (Or rejected, which is also not handled btw 🤔 )
**To Reproduce**
Steps to reproduce the behavior:
- [This codesandbox](https://codesandbox.io/s/envelop-async-schema-omsgc) demonstrates the issue by throwing in an artificial `setTimeout` of one second.
**Expected behavior**
1. When loading an async schema, no request should be handled before the schema is loaded
2. Rejection of the schema promise is handled gracefully
**Environment:**
- OS: macOS Monterey 12.0.1
- `@envelop/core`: latest
- NodeJS: 16.13.0
**Additional context**
I looked a bit into the codebase, here are some loose thoughts around how to achieve the first point I noted under "expected behavior":
- Making the `onPluginInit` hook async would easily solve this, but it would be a major change (atm the plugin initialisation and the call to `envelop` is sync). "Remembering" promises returned from that hook and awaiting them when calling `getEnveloped` would also be a breaking change.
- Maybe this could potentially even break other plugins that call `setSchema`, because it's non-deterministic in which order these calls will happen?
- I also found [this comment](https://github.com/dotansimha/envelop/discussions/497#discussioncomment-1063066) to be kinda related, but awaiting the schema to be loaded after calling `envelop` defeats the purpose of `useAsyncSchema` (you could just await before and use the sync `useSchema`)
- Maybe this can also be "solved" by more extensive documentation, i.e. "you need to make sure the schema promise is resolved before starting to handle request.
Would be open to help with this, but right now I'm not what the way forward would be 😄 | 1.0 | useAsyncSchema does not await the schema promise - **Describe the bug**
The `useAsyncSchema` accepts a promise that resolves to a GraphQL schema as argument. It chains a `.then()` on this promise (inside the call to `setSchema` happens), but since the envelop initialisation is synchronous, it might happen that the GraphQL server is handling the first request before the schema promise has resolved. (Or rejected, which is also not handled btw 🤔 )
**To Reproduce**
Steps to reproduce the behavior:
- [This codesandbox](https://codesandbox.io/s/envelop-async-schema-omsgc) demonstrates the issue by throwing in an artificial `setTimeout` of one second.
**Expected behavior**
1. When loading an async schema, no request should be handled before the schema is loaded
2. Rejection of the schema promise is handled gracefully
**Environment:**
- OS: macOS Monterey 12.0.1
- `@envelop/core`: latest
- NodeJS: 16.13.0
**Additional context**
I looked a bit into the codebase, here are some loose thoughts around how to achieve the first point I noted under "expected behavior":
- Making the `onPluginInit` hook async would easily solve this, but it would be a major change (atm the plugin initialisation and the call to `envelop` is sync). "Remembering" promises returned from that hook and awaiting them when calling `getEnveloped` would also be a breaking change.
- Maybe this could potentially even break other plugins that call `setSchema`, because it's non-deterministic in which order these calls will happen?
- I also found [this comment](https://github.com/dotansimha/envelop/discussions/497#discussioncomment-1063066) to be kinda related, but awaiting the schema to be loaded after calling `envelop` defeats the purpose of `useAsyncSchema` (you could just await before and use the sync `useSchema`)
- Maybe this can also be "solved" by more extensive documentation, i.e. "you need to make sure the schema promise is resolved before starting to handle request.
Would be open to help with this, but right now I'm not what the way forward would be 😄 | non_infrastructure | useasyncschema does not await the schema promise describe the bug the useasyncschema accepts a promise that resolves to a graphql schema as argument it chains a then on this promise inside the call to setschema happens but since the envelop initialisation is synchronous it might happen that the graphql server is handling the first request before the schema promise has resolved or rejected which is also not handled btw 🤔 to reproduce steps to reproduce the behavior demonstrates the issue by throwing in an artificial settimeout of one second expected behavior when loading an async schema no request should be handled before the schema is loaded rejection of the schema promise is handled gracefully environment os macos monterey envelop core latest nodejs additional context i looked a bit into the codebase here are some loose thoughts around how to achieve the first point i noted under expected behavior making the onplugininit hook async would easily solve this but it would be a major change atm the plugin initialisation and the call to envelop is sync remembering promises returned from that hook and awaiting them when calling getenveloped would also be a breaking change maybe this could potentially even break other plugins that call setschema because it s non deterministic in which order these calls will happen i also found to be kinda related but awaiting the schema to be loaded after calling envelop defeats the purpose of useasyncschema you could just await before and use the sync useschema maybe this can also be solved by more extensive documentation i e you need to make sure the schema promise is resolved before starting to handle request would be open to help with this but right now i m not what the way forward would be 😄 | 0 |
638,683 | 20,734,673,352 | IssuesEvent | 2022-03-14 12:42:31 | ooni/probe | https://api.github.com/repos/ooni/probe | opened | cli: experimental should be UnattendedOK | bug good first issue priority/medium ooni/probe-cli | This issue is about the CLI not running the experimental nettests group in unattended mode. We should actually be running this group in unattended mode. We investigated and it seems there's no particular reason why we ended up doing that, except perhaps an (undocumented) excess of caution. We can close this issue once we've marked experimental as UnattendedOK.
To this end, we need to head over to https://github.com/ooni/probe-cli/blob/master/cmd/ooniprobe/internal/nettests/groups.go and mark experimental as UnattendedOK: true. While there it would also be nice to mark performance explicitly as UnattendedOK: false. We should also probably document why that's the case (i.e., we don't want to max out the bandwidth with background tests).
As regards testing this functionality, it would probably suffice to pass the correct flag to `ooniprobe` to simulate running in unattended mode and manually verifying that the experimental group runs in this case. | 1.0 | cli: experimental should be UnattendedOK - This issue is about the CLI not running the experimental nettests group in unattended mode. We should actually be running this group in unattended mode. We investigated and it seems there's no particular reason why we ended up doing that, except perhaps an (undocumented) excess of caution. We can close this issue once we've marked experimental as UnattendedOK.
To this end, we need to head over to https://github.com/ooni/probe-cli/blob/master/cmd/ooniprobe/internal/nettests/groups.go and mark experimental as UnattendedOK: true. While there it would also be nice to mark performance explicitly as UnattendedOK: false. We should also probably document why that's the case (i.e., we don't want to max out the bandwidth with background tests).
As regards testing this functionality, it would probably suffice to pass the correct flag to `ooniprobe` to simulate running in unattended mode and manually verifying that the experimental group runs in this case. | non_infrastructure | cli experimental should be unattendedok this issue is about the cli not running the experimental nettests group in unattended mode we should actually be running this group in unattended mode we investigated and it seems there s no particular reason why we ended up doing that except perhaps an undocumented excess of caution we can close this issue once we ve marked experimental as unattendedok to this end we need to head over to and mark experimental as unattendedok true while there it would also be nice to mark performance explicitly as unattendedok false we should also probably document why that s the case i e we don t want to max out the bandwidth with background tests as regards testing this functionality it would probably suffice to pass the correct flag to ooniprobe to simulate running in unattended mode and manually verifying that the experimental group runs in this case | 0 |
484,553 | 13,941,700,607 | IssuesEvent | 2020-10-22 19:49:39 | neo-one-suite/neo-one | https://api.github.com/repos/neo-one-suite/neo-one | closed | Investigate/fix/update @neo-one/suite | kind/cleanup priority/important-longterm | As good as an idea as it was it actually doesn't really work in practice. When I tried to use it on deployment example repos using `@neo-one/suite` stopped automatic imports from working when you would expect it to. We should delete it and stop referring to it in the documentation just install the neo-one packages you are going to use. | 1.0 | Investigate/fix/update @neo-one/suite - As good as an idea as it was it actually doesn't really work in practice. When I tried to use it on deployment example repos using `@neo-one/suite` stopped automatic imports from working when you would expect it to. We should delete it and stop referring to it in the documentation just install the neo-one packages you are going to use. | non_infrastructure | investigate fix update neo one suite as good as an idea as it was it actually doesn t really work in practice when i tried to use it on deployment example repos using neo one suite stopped automatic imports from working when you would expect it to we should delete it and stop referring to it in the documentation just install the neo one packages you are going to use | 0 |
33,304 | 14,050,776,907 | IssuesEvent | 2020-11-02 12:19:30 | nestjs/nest | https://api.github.com/repos/nestjs/nest | closed | Can't send metadata with NestJS GRPC | scope: microservices type: question 🙌 | <!--
PLEASE HELP US PROCESS GITHUB ISSUES FASTER BY PROVIDING THE FOLLOWING INFORMATION.
ISSUES MISSING IMPORTANT INFORMATION MAY BE CLOSED WITHOUT INVESTIGATION.
-->
## I'm submitting a...
<!--
Please search GitHub for a similar issue or PR before submitting.
Check one of the following options with "x" -->
<pre><code>
[ ] Regression <!--(a behavior that used to work and stopped working in a new release)-->
[ ] Bug report
[x] Feature request
[ ] Documentation issue or request
[ ] Support request => Please do not submit support request here, instead post your question on Stack Overflow.
</code></pre>
## Current behavior
<!-- Describe how the issue manifests. -->
When I create a microservice GRPC server with NestJS, I can't send metadata in my response.
## Expected behavior
<!-- Describe what the desired behavior would be. -->
I want to be able to send headers to my client.
## Minimal reproduction of the problem with instructions
<!-- Please share a repo, a gist, or step-by-step instructions. -->
```
@GrpcMethod('AuthReadService', 'SignIn')
async signIn(
data: my.api.auth.SignInRequest,
): Promise<any> {
const response = await this.authService.signIn(data.email, data.password);
const metadata = new grpc.Metadata();
metadata.set(
'Set-Cookie',
`jwt=${response.jwt}; Expires=${response.expiresIn}; HttpOnly`,
);
// How can I send metadata to the client?
}
```
## What is the motivation / use case for changing the behavior?
<!-- Describe the motivation or the concrete use case. -->
I need to find a way to send the header `Set-Cookie` for authentication
## Environment
<pre><code>
Nest version: 6.0.0
<!-- Check whether this is still an issue in the most recent Nest version -->
For Tooling issues:
- Node version: v10.15.0 <!-- run `node --version` -->
- Platform: Mac <!-- Mac, Linux, Windows -->
Others:
<!-- Anything else relevant? Operating system version, IDE, package manager, ... -->
I'm using GRPC/Protobuf from browser with envoy proxy.
</code></pre> | 1.0 | Can't send metadata with NestJS GRPC - <!--
PLEASE HELP US PROCESS GITHUB ISSUES FASTER BY PROVIDING THE FOLLOWING INFORMATION.
ISSUES MISSING IMPORTANT INFORMATION MAY BE CLOSED WITHOUT INVESTIGATION.
-->
## I'm submitting a...
<!--
Please search GitHub for a similar issue or PR before submitting.
Check one of the following options with "x" -->
<pre><code>
[ ] Regression <!--(a behavior that used to work and stopped working in a new release)-->
[ ] Bug report
[x] Feature request
[ ] Documentation issue or request
[ ] Support request => Please do not submit support request here, instead post your question on Stack Overflow.
</code></pre>
## Current behavior
<!-- Describe how the issue manifests. -->
When I create a microservice GRPC server with NestJS, I can't send metadata in my response.
## Expected behavior
<!-- Describe what the desired behavior would be. -->
I want to be able to send headers to my client.
## Minimal reproduction of the problem with instructions
<!-- Please share a repo, a gist, or step-by-step instructions. -->
```
@GrpcMethod('AuthReadService', 'SignIn')
async signIn(
data: my.api.auth.SignInRequest,
): Promise<any> {
const response = await this.authService.signIn(data.email, data.password);
const metadata = new grpc.Metadata();
metadata.set(
'Set-Cookie',
`jwt=${response.jwt}; Expires=${response.expiresIn}; HttpOnly`,
);
// How can I send metadata to the client?
}
```
## What is the motivation / use case for changing the behavior?
<!-- Describe the motivation or the concrete use case. -->
I need to find a way to send the header `Set-Cookie` for authentication
## Environment
<pre><code>
Nest version: 6.0.0
<!-- Check whether this is still an issue in the most recent Nest version -->
For Tooling issues:
- Node version: v10.15.0 <!-- run `node --version` -->
- Platform: Mac <!-- Mac, Linux, Windows -->
Others:
<!-- Anything else relevant? Operating system version, IDE, package manager, ... -->
I'm using GRPC/Protobuf from browser with envoy proxy.
</code></pre> | non_infrastructure | can t send metadata with nestjs grpc please help us process github issues faster by providing the following information issues missing important information may be closed without investigation i m submitting a please search github for a similar issue or pr before submitting check one of the following options with x regression bug report feature request documentation issue or request support request please do not submit support request here instead post your question on stack overflow current behavior when i create a microservice grpc server with nestjs i can t send metadata in my response expected behavior i want to be able to send headers to my client minimal reproduction of the problem with instructions grpcmethod authreadservice signin async signin data my api auth signinrequest promise const response await this authservice signin data email data password const metadata new grpc metadata metadata set set cookie jwt response jwt expires response expiresin httponly how can i send metadata to the client what is the motivation use case for changing the behavior i need to find a way to send the header set cookie for authentication environment nest version for tooling issues node version platform mac others i m using grpc protobuf from browser with envoy proxy | 0 |
9,739 | 8,129,980,321 | IssuesEvent | 2018-08-17 16:49:30 | dotnet/corefx | https://api.github.com/repos/dotnet/corefx | closed | Start running CoreFX tests against Mono | area-Infrastructure test enhancement | Currently we run these tests against Desktop, Core, UWP on Windows, Unix, OSX where applicable. Now we have an increased focus on compatibility with Xamarin/Unity, we should consider having these tests run against Mono as well.
| 1.0 | Start running CoreFX tests against Mono - Currently we run these tests against Desktop, Core, UWP on Windows, Unix, OSX where applicable. Now we have an increased focus on compatibility with Xamarin/Unity, we should consider having these tests run against Mono as well.
| infrastructure | start running corefx tests against mono currently we run these tests against desktop core uwp on windows unix osx where applicable now we have an increased focus on compatibility with xamarin unity we should consider having these tests run against mono as well | 1 |
541,104 | 15,821,546,292 | IssuesEvent | 2021-04-05 20:42:25 | Rocket-Eagle/Rocket-Eagle | https://api.github.com/repos/Rocket-Eagle/Rocket-Eagle | closed | Add Backgroud Music Issue | Low Priority task | Add in background music
Technical Details:
- Issues related to this task are #28 | 1.0 | Add Backgroud Music Issue - Add in background music
Technical Details:
- Issues related to this task are #28 | non_infrastructure | add backgroud music issue add in background music technical details issues related to this task are | 0 |
235,857 | 19,432,313,301 | IssuesEvent | 2021-12-21 13:25:47 | pingcap/tidb | https://api.github.com/repos/pingcap/tidb | opened | Unstable test TestCacheTableBasicReadAndWrite | type/bug component/test severity/minor feature/developing | ## Bug Report
Please answer these questions before submitting your issue. Thanks!
### 1. Minimal reproduce step (Required)
<!-- a step by step guide for reproducing the bug. -->
Found in CI
https://ci.pingcap.net/blue/rest/organizations/jenkins/pipelines/tidb_ghpr_unit_test/runs/26002/nodes/52/steps/57/log/?start=0
```
[2021-12-20T08:11:05.302Z] --- FAIL: TestCacheTableBasicReadAndWrite (0.33s)
[2021-12-20T08:11:05.302Z] cache_test.go:225:
[2021-12-20T08:11:05.302Z] Error Trace: cache_test.go:225
[2021-12-20T08:11:05.302Z] Error: Should be true
[2021-12-20T08:11:05.302Z] Test: TestCacheTableBasicReadAndWrite
```
### 2. What did you expect to see? (Required)
CI stable
### 3. What did you see instead (Required)
Some time it fails
### 4. What is your TiDB version? (Required)
<!-- Paste the output of SELECT tidb_version() -->
Master | 1.0 | Unstable test TestCacheTableBasicReadAndWrite - ## Bug Report
Please answer these questions before submitting your issue. Thanks!
### 1. Minimal reproduce step (Required)
<!-- a step by step guide for reproducing the bug. -->
Found in CI
https://ci.pingcap.net/blue/rest/organizations/jenkins/pipelines/tidb_ghpr_unit_test/runs/26002/nodes/52/steps/57/log/?start=0
```
[2021-12-20T08:11:05.302Z] --- FAIL: TestCacheTableBasicReadAndWrite (0.33s)
[2021-12-20T08:11:05.302Z] cache_test.go:225:
[2021-12-20T08:11:05.302Z] Error Trace: cache_test.go:225
[2021-12-20T08:11:05.302Z] Error: Should be true
[2021-12-20T08:11:05.302Z] Test: TestCacheTableBasicReadAndWrite
```
### 2. What did you expect to see? (Required)
CI stable
### 3. What did you see instead (Required)
Some time it fails
### 4. What is your TiDB version? (Required)
<!-- Paste the output of SELECT tidb_version() -->
Master | non_infrastructure | unstable test testcachetablebasicreadandwrite bug report please answer these questions before submitting your issue thanks minimal reproduce step required found in ci fail testcachetablebasicreadandwrite cache test go error trace cache test go error should be true test testcachetablebasicreadandwrite what did you expect to see required ci stable what did you see instead required some time it fails what is your tidb version required master | 0 |
27,835 | 5,399,476,938 | IssuesEvent | 2017-02-27 19:33:09 | bounswe/bounswe2017group1 | https://api.github.com/repos/bounswe/bounswe2017group1 | closed | Failure Management section of the Requirements | documentation | Part 2.3.1 of the requirements is ambiguous. What are the appropriate messages or what are the possible errors? | 1.0 | Failure Management section of the Requirements - Part 2.3.1 of the requirements is ambiguous. What are the appropriate messages or what are the possible errors? | non_infrastructure | failure management section of the requirements part of the requirements is ambiguous what are the appropriate messages or what are the possible errors | 0 |
16,982 | 12,157,705,755 | IssuesEvent | 2020-04-25 23:31:17 | crystal-ball/crystal-ball.github.io | https://api.github.com/repos/crystal-ball/crystal-ball.github.io | closed | Chore: Automerge Renovate dev dependencies | Type: Infrastructure | Set Renovate to auto-merge dev dependency PRs where possible.
- Use PR title _'Reconfigure Renovate'_
- Include `"automerge": true` in dev deps group
## Updated
- [x] componentry.design
- [x] crystal-ball.github.io
- [x] react-application-prototype
- [x] node-service-prototype
- [x] babel-plugin-transform-import-aliases
- [x] componentry
- [x] dux-routing
- [x] svg-symbol-sprite-loader
- [x] babel-base
- [x] commitizen-base
- [x] eslint-config-eloquence
- [x] prettier-base
- [x] semantic-release base
- [x] webpack-base | 1.0 | Chore: Automerge Renovate dev dependencies - Set Renovate to auto-merge dev dependency PRs where possible.
- Use PR title _'Reconfigure Renovate'_
- Include `"automerge": true` in dev deps group
## Updated
- [x] componentry.design
- [x] crystal-ball.github.io
- [x] react-application-prototype
- [x] node-service-prototype
- [x] babel-plugin-transform-import-aliases
- [x] componentry
- [x] dux-routing
- [x] svg-symbol-sprite-loader
- [x] babel-base
- [x] commitizen-base
- [x] eslint-config-eloquence
- [x] prettier-base
- [x] semantic-release base
- [x] webpack-base | infrastructure | chore automerge renovate dev dependencies set renovate to auto merge dev dependency prs where possible use pr title reconfigure renovate include automerge true in dev deps group updated componentry design crystal ball github io react application prototype node service prototype babel plugin transform import aliases componentry dux routing svg symbol sprite loader babel base commitizen base eslint config eloquence prettier base semantic release base webpack base | 1 |
55,878 | 23,620,814,353 | IssuesEvent | 2022-08-24 20:21:04 | cityofaustin/atd-data-tech | https://api.github.com/repos/cityofaustin/atd-data-tech | closed | Reorder Inventory Report to match PR Tracker spreadsheet in TEST | Workgroup: AMD Service: Apps Type: Testing Product: Finance & Purchasing | After discovery meeting on 6/29. Talked to Brian V. and Ronnie about which fields they think they need in this new report page. Since had already prototyped a page in the Test environment, I was going to just adjust the fields in the system and add a few missing elements to make the table a bit better.
- [x] Reorder the fields to be in the same order as PR Tracker spreadsheet
- [ ] Add `Purchase Type` - text formula to `purchase_request_items`
- ended up using the connected field but might go back to this if I can't control the display rules on the table without these fields
- [ ] Add `Purchase ID` - text formula to `purchase_request_items`
- ended up using the connected field but might go back to this if I can't control the display rules on the table without these fields | 1.0 | Reorder Inventory Report to match PR Tracker spreadsheet in TEST - After discovery meeting on 6/29. Talked to Brian V. and Ronnie about which fields they think they need in this new report page. Since had already prototyped a page in the Test environment, I was going to just adjust the fields in the system and add a few missing elements to make the table a bit better.
- [x] Reorder the fields to be in the same order as PR Tracker spreadsheet
- [ ] Add `Purchase Type` - text formula to `purchase_request_items`
- ended up using the connected field but might go back to this if I can't control the display rules on the table without these fields
- [ ] Add `Purchase ID` - text formula to `purchase_request_items`
- ended up using the connected field but might go back to this if I can't control the display rules on the table without these fields | non_infrastructure | reorder inventory report to match pr tracker spreadsheet in test after discovery meeting on talked to brian v and ronnie about which fields they think they need in this new report page since had already prototyped a page in the test environment i was going to just adjust the fields in the system and add a few missing elements to make the table a bit better reorder the fields to be in the same order as pr tracker spreadsheet add purchase type text formula to purchase request items ended up using the connected field but might go back to this if i can t control the display rules on the table without these fields add purchase id text formula to purchase request items ended up using the connected field but might go back to this if i can t control the display rules on the table without these fields | 0 |
21,592 | 7,047,279,366 | IssuesEvent | 2018-01-02 12:42:42 | ShaikASK/Testing | https://api.github.com/repos/ShaikASK/Testing | closed | Signup::Duplicate emails are triggered upon signing up | AskIT Build Version #2 Defect P2 | Steps To Replicate :
1. Launch the url : http://192.168.1.197:9090/#/signup
2. Enter all the valid information in the fields
3. Click on save button
4. "Thanks you" page is displayed
5. Check the email configured in the system.
Experienced Behavior: Observed that duplicate emails are triggered and a part of notification
Expected Behavior: Ensure that there only one email should be triggered as a part of notification
| 1.0 | Signup::Duplicate emails are triggered upon signing up - Steps To Replicate :
1. Launch the url : http://192.168.1.197:9090/#/signup
2. Enter all the valid information in the fields
3. Click on save button
4. "Thanks you" page is displayed
5. Check the email configured in the system.
Experienced Behavior: Observed that duplicate emails are triggered and a part of notification
Expected Behavior: Ensure that there only one email should be triggered as a part of notification
| non_infrastructure | signup duplicate emails are triggered upon signing up steps to replicate launch the url enter all the valid information in the fields click on save button thanks you page is displayed check the email configured in the system experienced behavior observed that duplicate emails are triggered and a part of notification expected behavior ensure that there only one email should be triggered as a part of notification | 0 |
7,374 | 6,926,070,964 | IssuesEvent | 2017-11-30 17:52:25 | Azure/azure-cli | https://api.github.com/repos/Azure/azure-cli | opened | Generic Update needs to support custom setter and getter | Infrastructure Knack | Most generic updates use the SDK for the getter and setter and a custom method for the convenience arguments, so the Knack generic update was set up like that. However, there are situations where the getter and/or setter come from a custom source, so it needs to support that. | 1.0 | Generic Update needs to support custom setter and getter - Most generic updates use the SDK for the getter and setter and a custom method for the convenience arguments, so the Knack generic update was set up like that. However, there are situations where the getter and/or setter come from a custom source, so it needs to support that. | infrastructure | generic update needs to support custom setter and getter most generic updates use the sdk for the getter and setter and a custom method for the convenience arguments so the knack generic update was set up like that however there are situations where the getter and or setter come from a custom source so it needs to support that | 1 |
15,671 | 11,647,611,552 | IssuesEvent | 2020-03-01 16:05:52 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | AzDO losing connection to build agents | area-Infrastructure blocking-clean-ci untriaged | Build jobs are failing with the following message
> We stopped hearing from agent NetCorePublic-Pool 93. Verify the agent machine is running and has a healthy network connection. Anything that terminates an agent process, starves it for CPU, or blocks its network access can cause this error. For more information, see: https://go.microsoft.com/fwlink/?linkid=846610
Impact over last 100 builds
|Build|Kind|Timeline Record|
|---|---|---|
|[540296](https://dev.azure.com/dnceng/public/_build/results?buildId=540296)|PR https://github.com/dotnet/runtime/pull/32858|CoreCLR Pri0 Test Run Windows_NT x64 checked|
|[540296](https://dev.azure.com/dnceng/public/_build/results?buildId=540296)|PR https://github.com/dotnet/runtime/pull/32858|CoreCLR Pri0 Test Run Linux x64 checked|
|[540296](https://dev.azure.com/dnceng/public/_build/results?buildId=540296)|PR https://github.com/dotnet/runtime/pull/32858|Test crossgen-comparison Linux arm checked|
|[540296](https://dev.azure.com/dnceng/public/_build/results?buildId=540296)|PR https://github.com/dotnet/runtime/pull/32858|CoreCLR Pri0 Test Build Linux x64 checked|
|[540296](https://dev.azure.com/dnceng/public/_build/results?buildId=540296)|PR https://github.com/dotnet/runtime/pull/32858|CoreCLR Pri0 Test Run Windows_NT x86 checked|
|[540296](https://dev.azure.com/dnceng/public/_build/results?buildId=540296)|PR https://github.com/dotnet/runtime/pull/32858|Installer Build and Test Linux_arm64 Debug|
|[540296](https://dev.azure.com/dnceng/public/_build/results?buildId=540296)|PR https://github.com/dotnet/runtime/pull/32858|Installer Build and Test Linux_musl_x64 Debug|
|[540296](https://dev.azure.com/dnceng/public/_build/results?buildId=540296)|PR https://github.com/dotnet/runtime/pull/32858|Libraries Test Build Linux x64 Debug|
|[540296](https://dev.azure.com/dnceng/public/_build/results?buildId=540296)|PR https://github.com/dotnet/runtime/pull/32858|Installer Build and Test Linux_arm Debug|
|[540293](https://dev.azure.com/dnceng/public/_build/results?buildId=540293)|PR https://github.com/dotnet/runtime/pull/32858|CoreCLR Pri0 Test Build Linux x64 checked|
|[540293](https://dev.azure.com/dnceng/public/_build/results?buildId=540293)|PR https://github.com/dotnet/runtime/pull/32858|CoreCLR Pri0 Test Run Windows_NT x64 checked|
|[540293](https://dev.azure.com/dnceng/public/_build/results?buildId=540293)|PR https://github.com/dotnet/runtime/pull/32858|CoreCLR Pri0 Test Run Windows_NT x86 checked|
|[540293](https://dev.azure.com/dnceng/public/_build/results?buildId=540293)|PR https://github.com/dotnet/runtime/pull/32858|Libraries Test Build Linux x64 Debug|
|[540293](https://dev.azure.com/dnceng/public/_build/results?buildId=540293)|PR https://github.com/dotnet/runtime/pull/32858|Installer Build and Test Linux_arm Debug|
|[540233](https://dev.azure.com/dnceng/public/_build/results?buildId=540233)|Rolling|CoreCLR Pri0 Test Run Linux arm64 checked|
|[540233](https://dev.azure.com/dnceng/public/_build/results?buildId=540233)|Rolling|Libraries Test Run release coreclr Linux x64 Release|
|[540233](https://dev.azure.com/dnceng/public/_build/results?buildId=540233)|Rolling|Libraries Test Run release mono Linux x64 Release|
|[540233](https://dev.azure.com/dnceng/public/_build/results?buildId=540233)|Rolling|Libraries Test Run release coreclr Linux arm Release|
|[540233](https://dev.azure.com/dnceng/public/_build/results?buildId=540233)|Rolling|Libraries Test Run release coreclr Linux_musl x64 Release|
|[540233](https://dev.azure.com/dnceng/public/_build/results?buildId=540233)|Rolling|CoreCLR Pri0 Test Run Linux arm checked|
|[540233](https://dev.azure.com/dnceng/public/_build/results?buildId=540233)|Rolling|Installer Build and Test Linux_arm Release| | 1.0 | AzDO losing connection to build agents - Build jobs are failing with the following message
> We stopped hearing from agent NetCorePublic-Pool 93. Verify the agent machine is running and has a healthy network connection. Anything that terminates an agent process, starves it for CPU, or blocks its network access can cause this error. For more information, see: https://go.microsoft.com/fwlink/?linkid=846610
Impact over last 100 builds
|Build|Kind|Timeline Record|
|---|---|---|
|[540296](https://dev.azure.com/dnceng/public/_build/results?buildId=540296)|PR https://github.com/dotnet/runtime/pull/32858|CoreCLR Pri0 Test Run Windows_NT x64 checked|
|[540296](https://dev.azure.com/dnceng/public/_build/results?buildId=540296)|PR https://github.com/dotnet/runtime/pull/32858|CoreCLR Pri0 Test Run Linux x64 checked|
|[540296](https://dev.azure.com/dnceng/public/_build/results?buildId=540296)|PR https://github.com/dotnet/runtime/pull/32858|Test crossgen-comparison Linux arm checked|
|[540296](https://dev.azure.com/dnceng/public/_build/results?buildId=540296)|PR https://github.com/dotnet/runtime/pull/32858|CoreCLR Pri0 Test Build Linux x64 checked|
|[540296](https://dev.azure.com/dnceng/public/_build/results?buildId=540296)|PR https://github.com/dotnet/runtime/pull/32858|CoreCLR Pri0 Test Run Windows_NT x86 checked|
|[540296](https://dev.azure.com/dnceng/public/_build/results?buildId=540296)|PR https://github.com/dotnet/runtime/pull/32858|Installer Build and Test Linux_arm64 Debug|
|[540296](https://dev.azure.com/dnceng/public/_build/results?buildId=540296)|PR https://github.com/dotnet/runtime/pull/32858|Installer Build and Test Linux_musl_x64 Debug|
|[540296](https://dev.azure.com/dnceng/public/_build/results?buildId=540296)|PR https://github.com/dotnet/runtime/pull/32858|Libraries Test Build Linux x64 Debug|
|[540296](https://dev.azure.com/dnceng/public/_build/results?buildId=540296)|PR https://github.com/dotnet/runtime/pull/32858|Installer Build and Test Linux_arm Debug|
|[540293](https://dev.azure.com/dnceng/public/_build/results?buildId=540293)|PR https://github.com/dotnet/runtime/pull/32858|CoreCLR Pri0 Test Build Linux x64 checked|
|[540293](https://dev.azure.com/dnceng/public/_build/results?buildId=540293)|PR https://github.com/dotnet/runtime/pull/32858|CoreCLR Pri0 Test Run Windows_NT x64 checked|
|[540293](https://dev.azure.com/dnceng/public/_build/results?buildId=540293)|PR https://github.com/dotnet/runtime/pull/32858|CoreCLR Pri0 Test Run Windows_NT x86 checked|
|[540293](https://dev.azure.com/dnceng/public/_build/results?buildId=540293)|PR https://github.com/dotnet/runtime/pull/32858|Libraries Test Build Linux x64 Debug|
|[540293](https://dev.azure.com/dnceng/public/_build/results?buildId=540293)|PR https://github.com/dotnet/runtime/pull/32858|Installer Build and Test Linux_arm Debug|
|[540233](https://dev.azure.com/dnceng/public/_build/results?buildId=540233)|Rolling|CoreCLR Pri0 Test Run Linux arm64 checked|
|[540233](https://dev.azure.com/dnceng/public/_build/results?buildId=540233)|Rolling|Libraries Test Run release coreclr Linux x64 Release|
|[540233](https://dev.azure.com/dnceng/public/_build/results?buildId=540233)|Rolling|Libraries Test Run release mono Linux x64 Release|
|[540233](https://dev.azure.com/dnceng/public/_build/results?buildId=540233)|Rolling|Libraries Test Run release coreclr Linux arm Release|
|[540233](https://dev.azure.com/dnceng/public/_build/results?buildId=540233)|Rolling|Libraries Test Run release coreclr Linux_musl x64 Release|
|[540233](https://dev.azure.com/dnceng/public/_build/results?buildId=540233)|Rolling|CoreCLR Pri0 Test Run Linux arm checked|
|[540233](https://dev.azure.com/dnceng/public/_build/results?buildId=540233)|Rolling|Installer Build and Test Linux_arm Release| | infrastructure | azdo losing connection to build agents build jobs are failing with the following message we stopped hearing from agent netcorepublic pool verify the agent machine is running and has a healthy network connection anything that terminates an agent process starves it for cpu or blocks its network access can cause this error for more information see impact over last builds build kind timeline record test run windows nt checked test run linux checked crossgen comparison linux arm checked test build linux checked test run windows nt checked build and test linux debug build and test linux musl debug test build linux debug build and test linux arm debug test build linux checked test run windows nt checked test run windows nt checked test build linux debug build and test linux arm debug test run linux checked test run release coreclr linux release test run release mono linux release test run release coreclr linux arm release test run release coreclr linux musl release test run linux arm checked build and test linux arm release | 1 |
18,813 | 11,081,178,690 | IssuesEvent | 2019-12-13 09:13:20 | kyma-project/kyma | https://api.github.com/repos/kyma-project/kyma | opened | Persisance profile for ory components | area/service-mesh enhancement | <!-- Thank you for your contribution. Before you submit the issue:
1. Search open and closed issues for duplicates.
2. Read the contributing guidelines.
-->
**Description**
Currently ory components like hydra does not have any persistance. If you create an Oauth2 client hydra is storing it locally on container disk. If application restarts information about clients is lost.
We need to have an easily configurable profile (helm override) that will setup ory components with persistence enabled.
The story is with oathkeeper.
<!-- Provide a clear and concise description of the feature. -->
**Reasons**
In memory or basically local storage is good for playing around but not for the production use case
<!-- Explain why we should add this feature. Provide use cases to illustrate its benefits. -->
**Attachments**
<!-- Attach any files, links, code samples, or screenshots that will convince us to your idea. -->
| 1.0 | Persisance profile for ory components - <!-- Thank you for your contribution. Before you submit the issue:
1. Search open and closed issues for duplicates.
2. Read the contributing guidelines.
-->
**Description**
Currently ory components like hydra does not have any persistance. If you create an Oauth2 client hydra is storing it locally on container disk. If application restarts information about clients is lost.
We need to have an easily configurable profile (helm override) that will setup ory components with persistence enabled.
The story is with oathkeeper.
<!-- Provide a clear and concise description of the feature. -->
**Reasons**
In memory or basically local storage is good for playing around but not for the production use case
<!-- Explain why we should add this feature. Provide use cases to illustrate its benefits. -->
**Attachments**
<!-- Attach any files, links, code samples, or screenshots that will convince us to your idea. -->
| non_infrastructure | persisance profile for ory components thank you for your contribution before you submit the issue search open and closed issues for duplicates read the contributing guidelines description currently ory components like hydra does not have any persistance if you create an client hydra is storing it locally on container disk if application restarts information about clients is lost we need to have an easily configurable profile helm override that will setup ory components with persistence enabled the story is with oathkeeper reasons in memory or basically local storage is good for playing around but not for the production use case attachments | 0 |
220,675 | 24,565,373,052 | IssuesEvent | 2022-10-13 02:09:31 | dgee2/dgee2.github.io | https://api.github.com/repos/dgee2/dgee2.github.io | closed | CVE-2022-2900 (High) detected in parse-url-5.0.2.tgz - autoclosed | security vulnerability | ## CVE-2022-2900 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>parse-url-5.0.2.tgz</b></p></summary>
<p>An advanced url parser supporting git urls too.</p>
<p>Library home page: <a href="https://registry.npmjs.org/parse-url/-/parse-url-5.0.2.tgz">https://registry.npmjs.org/parse-url/-/parse-url-5.0.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/parse-url/package.json</p>
<p>
Dependency Hierarchy:
- gatsby-3.2.1.tgz (Root Library)
- gatsby-telemetry-2.2.0.tgz
- git-up-4.0.2.tgz
- :x: **parse-url-5.0.2.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Server-Side Request Forgery (SSRF) in GitHub repository ionicabizau/parse-url prior to 8.1.0.
<p>Publish Date: 2022-09-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-2900>CVE-2022-2900</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-09-14</p>
<p>Fix Resolution (parse-url): 8.0.0</p>
<p>Direct dependency fix Resolution (gatsby): 4.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-2900 (High) detected in parse-url-5.0.2.tgz - autoclosed - ## CVE-2022-2900 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>parse-url-5.0.2.tgz</b></p></summary>
<p>An advanced url parser supporting git urls too.</p>
<p>Library home page: <a href="https://registry.npmjs.org/parse-url/-/parse-url-5.0.2.tgz">https://registry.npmjs.org/parse-url/-/parse-url-5.0.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/parse-url/package.json</p>
<p>
Dependency Hierarchy:
- gatsby-3.2.1.tgz (Root Library)
- gatsby-telemetry-2.2.0.tgz
- git-up-4.0.2.tgz
- :x: **parse-url-5.0.2.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Server-Side Request Forgery (SSRF) in GitHub repository ionicabizau/parse-url prior to 8.1.0.
<p>Publish Date: 2022-09-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-2900>CVE-2022-2900</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-09-14</p>
<p>Fix Resolution (parse-url): 8.0.0</p>
<p>Direct dependency fix Resolution (gatsby): 4.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | cve high detected in parse url tgz autoclosed cve high severity vulnerability vulnerable library parse url tgz an advanced url parser supporting git urls too library home page a href path to dependency file package json path to vulnerable library node modules parse url package json dependency hierarchy gatsby tgz root library gatsby telemetry tgz git up tgz x parse url tgz vulnerable library found in base branch master vulnerability details server side request forgery ssrf in github repository ionicabizau parse url prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution parse url direct dependency fix resolution gatsby step up your open source security game with mend | 0 |
34,347 | 29,508,186,197 | IssuesEvent | 2023-06-03 15:27:21 | norswap/0xFable | https://api.github.com/repos/norswap/0xFable | opened | Spin up own chain infrastructure | required open-ended infrastructure | For the tech demo, the game has to run somewhere. The long-term plan is to make it a rollup (settlement layer TBD), but in the short term, we want an easier solution.
For cost considerations, the easiest thing to do would be to spin our own Proof-of-Authoriy (PoA) chain. This would also enable to very easily make the game totally free [via a burner wallet](https://github.com/norswap/0xFable/issues/31). | 1.0 | Spin up own chain infrastructure - For the tech demo, the game has to run somewhere. The long-term plan is to make it a rollup (settlement layer TBD), but in the short term, we want an easier solution.
For cost considerations, the easiest thing to do would be to spin our own Proof-of-Authoriy (PoA) chain. This would also enable to very easily make the game totally free [via a burner wallet](https://github.com/norswap/0xFable/issues/31). | infrastructure | spin up own chain infrastructure for the tech demo the game has to run somewhere the long term plan is to make it a rollup settlement layer tbd but in the short term we want an easier solution for cost considerations the easiest thing to do would be to spin our own proof of authoriy poa chain this would also enable to very easily make the game totally free | 1 |
9,442 | 2,615,150,320 | IssuesEvent | 2015-03-01 06:27:35 | chrsmith/reaver-wps | https://api.github.com/repos/chrsmith/reaver-wps | closed | Wash not installing with Reaver via apt-get | auto-migrated Priority-Triage Type-Defect | ```
Hey guys, thanks so much for Reaver, I really appreciate the effort used to
make this tool. I noticed that when I try to install reaver in Backtrack 5 R1
using apt-get, wash is not included. Could wash be uploaded to the apt-get
repos separately? Or perhaps an updated reaver package?
Thanks!
0. What version of Reaver are you using? (Only defects against the latest
version will be considered.)
1.3
1. What operating system are you using (Linux is the only supported OS)?
Backtrack 5 R1
5. What is the entire command line string you are supplying to reaver?
apt-get install reaver
6. Please describe what you think the issue is.
needs to be updated
```
Original issue reported on code.google.com by `joshuavi...@gmail.com` on 17 Jan 2012 at 5:51 | 1.0 | Wash not installing with Reaver via apt-get - ```
Hey guys, thanks so much for Reaver, I really appreciate the effort used to
make this tool. I noticed that when I try to install reaver in Backtrack 5 R1
using apt-get, wash is not included. Could wash be uploaded to the apt-get
repos separately? Or perhaps an updated reaver package?
Thanks!
0. What version of Reaver are you using? (Only defects against the latest
version will be considered.)
1.3
1. What operating system are you using (Linux is the only supported OS)?
Backtrack 5 R1
5. What is the entire command line string you are supplying to reaver?
apt-get install reaver
6. Please describe what you think the issue is.
needs to be updated
```
Original issue reported on code.google.com by `joshuavi...@gmail.com` on 17 Jan 2012 at 5:51 | non_infrastructure | wash not installing with reaver via apt get hey guys thanks so much for reaver i really appreciate the effort used to make this tool i noticed that when i try to install reaver in backtrack using apt get wash is not included could wash be uploaded to the apt get repos separately or perhaps an updated reaver package thanks what version of reaver are you using only defects against the latest version will be considered what operating system are you using linux is the only supported os backtrack what is the entire command line string you are supplying to reaver apt get install reaver please describe what you think the issue is needs to be updated original issue reported on code google com by joshuavi gmail com on jan at | 0 |
261,259 | 22,710,035,817 | IssuesEvent | 2022-07-05 18:19:50 | saltstack/salt | https://api.github.com/repos/saltstack/salt | closed | [BUG] postgres_privileges: unable to verify existing privileges on views. | Execution-Module Bug severity-medium help-wanted Needs-Testcase | Hello !
`postgres.privileges_list` and similar functions are unable to verify existing privileges on views.
`postgres_privileges.present` says:
> View permissions should specify `object_type: table`.
https://github.com/saltstack/salt/blob/a670b4ae72ec11f5485c216c54059e14223019b8/salt/states/postgres_privileges.py#L118
But `_make_privileges_list_query` is unable to verify existing privileges on views because it filters on ordinary tables only.
https://github.com/saltstack/salt/blob/f41cda17ce6cc0ee54df483f4d5f48cc2b9632e7/salt/modules/postgres.py#L2728
A little update should fix that.
```diff
- "AND relkind = 'r'",
+ "AND relkind in ('r', 'v')",
```
State example : `pg_catalog.pg_shadow` is a view.
```yaml
pgbouncer_auth_user:
postgres_privileges.present:
# GRANT SELECT ON pg_catalog.pg_shadow TO pgbouncer_auth_user;
- prepend: pg_catalog
- object_name: pg_shadow
- object_type: table
- privileges:
- SELECT
- maintenance_db: my_db
- db_port: 5432
```
| 1.0 | [BUG] postgres_privileges: unable to verify existing privileges on views. - Hello !
`postgres.privileges_list` and similar functions are unable to verify existing privileges on views.
`postgres_privileges.present` says:
> View permissions should specify `object_type: table`.
https://github.com/saltstack/salt/blob/a670b4ae72ec11f5485c216c54059e14223019b8/salt/states/postgres_privileges.py#L118
But `_make_privileges_list_query` is unable to verify existing privileges on views because it filters on ordinary tables only.
https://github.com/saltstack/salt/blob/f41cda17ce6cc0ee54df483f4d5f48cc2b9632e7/salt/modules/postgres.py#L2728
A little update should fix that.
```diff
- "AND relkind = 'r'",
+ "AND relkind in ('r', 'v')",
```
State example : `pg_catalog.pg_shadow` is a view.
```yaml
pgbouncer_auth_user:
postgres_privileges.present:
# GRANT SELECT ON pg_catalog.pg_shadow TO pgbouncer_auth_user;
- prepend: pg_catalog
- object_name: pg_shadow
- object_type: table
- privileges:
- SELECT
- maintenance_db: my_db
- db_port: 5432
```
| non_infrastructure | postgres privileges unable to verify existing privileges on views hello postgres privileges list and similar functions are unable to verify existing privileges on views postgres privileges present says view permissions should specify object type table but make privileges list query is unable to verify existing privileges on views because it filters on ordinary tables only a little update should fix that diff and relkind r and relkind in r v state example pg catalog pg shadow is a view yaml pgbouncer auth user postgres privileges present grant select on pg catalog pg shadow to pgbouncer auth user prepend pg catalog object name pg shadow object type table privileges select maintenance db my db db port | 0 |
25,867 | 19,305,478,387 | IssuesEvent | 2021-12-13 11:02:41 | arduino/arduino-ide | https://api.github.com/repos/arduino/arduino-ide | opened | Check for changes to formatter output resulting from clangd bump | type: enhancement topic: infrastructure topic: language server | ## Describe the current behavior
The Arduino IDE's **Tools > Auto Format** functionality is provided by [ClangFormat](https://clang.llvm.org/docs/ClangFormat.html) (via [clangd](https://clangd.llvm.org/) and [Arduino Language Server](https://github.com/arduino/arduino-language-server)).
Arduino Language Server contains an embedded ClangFormat configuration that defines the standard Arduino code formatting style, which is used as the default Auto Format configuration:
https://github.com/arduino/arduino-language-server/blob/0.5.0/ls/ls_formatter.go#L15-L159
This configuration was developed using [ClangFormat 11.0.1](https://releases.llvm.org/11.0.1/tools/clang/docs/ClangFormatStyleOptions.html), but will be used with whatever version of clangd is installed with the Arduino IDE (currently ~12.0.0).
We received valued advice from someone with experience using ClangFormat on Arduino code:
https://github.com/arduino/Arduino/pull/11543#issuecomment-850080301
> we use clang-format-8 and believe me we have learned they are not all the same so you really have to pick one and stick to it. that said, if y'all want to pick one clang format version i'm happy to change our CI to match
ClangFormat has a strict approach to formatting. While the formatting style is very configurable, often it is not possible to configure it to leave the code as-is. This means that newly introduced configurations are likely to have a default setting that imposes formatting of some form, with no guarantees that it will align with the official Arduino code style.
## To reproduce
1. Start the Arduino IDE.
1. Open a sketch that does not [contain a `.clang-format` file](https://zube.io/arduino/tooling-team/c/9160).
1. Select **Tools > Auto Format** from the Arduino IDE menus.
The sketch file currently open in the editor will be formatted according to the ClangFormat configuration embedded in Arduino Language Server.
## Describe the request
Set up a formal system to check for formatter output changes at every update to the Arduino IDE 2.x clangd dependency:
https://github.com/arduino/arduino-ide/blob/2.0.0-rc2/arduino-ide-extension/scripts/download-ls.js#L8
My proposal is that we produce a file containing test data code that will exercise the significant C++ formatting capabilities of ClangFormat then check for a diff after formatting that code with the new version. If there is no diff, then we have a reasonable certainty that the bump will not necessitate any adjustments to the ClangFormat configuration.
Ideally this would be set up to run automatically as part of the CI/CD system of the appropriate repository. Since the clangd version in use is currently defined in this repository, it seems to be the best place.
## Additional context
Related:
- https://github.com/arduino/arduino-ide/issues/42 | 1.0 | Check for changes to formatter output resulting from clangd bump - ## Describe the current behavior
The Arduino IDE's **Tools > Auto Format** functionality is provided by [ClangFormat](https://clang.llvm.org/docs/ClangFormat.html) (via [clangd](https://clangd.llvm.org/) and [Arduino Language Server](https://github.com/arduino/arduino-language-server)).
Arduino Language Server contains an embedded ClangFormat configuration that defines the standard Arduino code formatting style, which is used as the default Auto Format configuration:
https://github.com/arduino/arduino-language-server/blob/0.5.0/ls/ls_formatter.go#L15-L159
This configuration was developed using [ClangFormat 11.0.1](https://releases.llvm.org/11.0.1/tools/clang/docs/ClangFormatStyleOptions.html), but will be used with whatever version of clangd is installed with the Arduino IDE (currently ~12.0.0).
We received valued advice from someone with experience using ClangFormat on Arduino code:
https://github.com/arduino/Arduino/pull/11543#issuecomment-850080301
> we use clang-format-8 and believe me we have learned they are not all the same so you really have to pick one and stick to it. that said, if y'all want to pick one clang format version i'm happy to change our CI to match
ClangFormat has a strict approach to formatting. While the formatting style is very configurable, often it is not possible to configure it to leave the code as-is. This means that newly introduced configurations are likely to have a default setting that imposes formatting of some form, with no guarantees that it will align with the official Arduino code style.
## To reproduce
1. Start the Arduino IDE.
1. Open a sketch that does not [contain a `.clang-format` file](https://zube.io/arduino/tooling-team/c/9160).
1. Select **Tools > Auto Format** from the Arduino IDE menus.
The sketch file currently open in the editor will be formatted according to the ClangFormat configuration embedded in Arduino Language Server.
## Describe the request
Set up a formal system to check for formatter output changes at every update to the Arduino IDE 2.x clangd dependency:
https://github.com/arduino/arduino-ide/blob/2.0.0-rc2/arduino-ide-extension/scripts/download-ls.js#L8
My proposal is that we produce a file containing test data code that will exercise the significant C++ formatting capabilities of ClangFormat then check for a diff after formatting that code with the new version. If there is no diff, then we have a reasonable certainty that the bump will not necessitate any adjustments to the ClangFormat configuration.
Ideally this would be set up to run automatically as part of the CI/CD system of the appropriate repository. Since the clangd version in use is currently defined in this repository, it seems to be the best place.
## Additional context
Related:
- https://github.com/arduino/arduino-ide/issues/42 | infrastructure | check for changes to formatter output resulting from clangd bump describe the current behavior the arduino ide s tools auto format functionality is provided by via and arduino language server contains an embedded clangformat configuration that defines the standard arduino code formatting style which is used as the default auto format configuration this configuration was developed using but will be used with whatever version of clangd is installed with the arduino ide currently we received valued advice from someone with experience using clangformat on arduino code we use clang format and believe me we have learned they are not all the same so you really have to pick one and stick to it that said if y all want to pick one clang format version i m happy to change our ci to match clangformat has a strict approach to formatting while the formatting style is very configurable often it is not possible to configure it to leave the code as is this means that newly introduced configurations are likely to have a default setting that imposes formatting of some form with no guarantees that it will align with the official arduino code style to reproduce start the arduino ide open a sketch that does not select tools auto format from the arduino ide menus the sketch file currently open in the editor will be formatted according to the clangformat configuration embedded in arduino language server describe the request set up a formal system to check for formatter output changes at every update to the arduino ide x clangd dependency my proposal is that we produce a file containing test data code that will exercise the significant c formatting capabilities of clangformat then check for a diff after formatting that code with the new version if there is no diff then we have a reasonable certainty that the bump will not necessitate any adjustments to the clangformat configuration ideally this would be set up to run automatically as part of the ci cd system of the appropriate repository since the clangd version in use is currently defined in this repository it seems to be the best place additional context related | 1 |
2,773 | 3,863,145,056 | IssuesEvent | 2016-04-08 08:03:51 | ideoforms/signal | https://api.github.com/repos/ideoforms/signal | opened | Further examples | infrastructure | JSON load/save
`FFT`
`AudioIn`
`Supersaw` (multichannel expansion)
Vamp segmentation
`Waveshaper`
`Index`/`Counter`
| 1.0 | Further examples - JSON load/save
`FFT`
`AudioIn`
`Supersaw` (multichannel expansion)
Vamp segmentation
`Waveshaper`
`Index`/`Counter`
| infrastructure | further examples json load save fft audioin supersaw multichannel expansion vamp segmentation waveshaper index counter | 1 |
628,087 | 19,975,233,882 | IssuesEvent | 2022-01-29 01:46:12 | Baystation12/Baystation12 | https://api.github.com/repos/Baystation12/Baystation12 | closed | Mapped cargo tug self-links incorrectly | Bug :bug: Priority: Trivial Could Reproduce :bug: | #### Description of issue
The cargo tug and train mapped into the cargo warehouse do not link themselves correctly when initialized. The central trolley hitches itself both to the cargo engine and the third trolley, while the engine and trailing trolley both hitch themselves to the central trolley, resulting in a tangled mess of A<->B<->C when the code expected A<-B<-C. This in turn causes the tug to not move and instead eject a user when they try to drive it.
#### Difference between expected and actual behavior
Engine and trolleys should correctly chain themselves instead of creating broken cross-links.
#### Workaround
Use the right-click menu to unhitch the engine and both trolleys, then reconnect everything together with click+drag. This will allow the tug to be driveable again.
#### Steps to reproduce
1. Start the Torch map
2. Go to the cargo warehouse
3. Try to use the cargo tug
#### Specific information for locating
Engine: `/obj/vehicle/train/cargo/engine`
Trolley: `/obj/vehicle/train/cargo/trolley`
Very likely an issue with the base `/obj/vehicle/train` type.
#### Length of time in which bug has been known to occur
Unknown, reported today in-game, stated to have been going on 'forever'.
#### Client version, Server revision & Game ID
Client Version: 513
Server Revision: 02aeb12d1f76e63b403f40b988cd8450ba17ec5a - dev - 2021-07-21
Game ID: cda-ahmg
Current map: SEV Torch
#### Issue bingo
<!-- Check these by writing an x inside the [ ] (like this: [x])-->
<!-- Don't forget to remove the space between the brackets, or it won't work! -->
- [X] Issue could be reproduced at least once
- [X] Issue could be reproduced by different players
- [X] Issue could be reproduced in multiple rounds
- [X] Issue happened in a recent (less than 7 days ago) round
- [X] [Couldn't find an existing issue about this](https://github.com/Baystation12/Baystation12/issues)
| 1.0 | Mapped cargo tug self-links incorrectly - #### Description of issue
The cargo tug and train mapped into the cargo warehouse do not link themselves correctly when initialized. The central trolley hitches itself both to the cargo engine and the third trolley, while the engine and trailing trolley both hitch themselves to the central trolley, resulting in a tangled mess of A<->B<->C when the code expected A<-B<-C. This in turn causes the tug to not move and instead eject a user when they try to drive it.
#### Difference between expected and actual behavior
Engine and trolleys should correctly chain themselves instead of creating broken cross-links.
#### Workaround
Use the right-click menu to unhitch the engine and both trolleys, then reconnect everything together with click+drag. This will allow the tug to be driveable again.
#### Steps to reproduce
1. Start the Torch map
2. Go to the cargo warehouse
3. Try to use the cargo tug
#### Specific information for locating
Engine: `/obj/vehicle/train/cargo/engine`
Trolley: `/obj/vehicle/train/cargo/trolley`
Very likely an issue with the base `/obj/vehicle/train` type.
#### Length of time in which bug has been known to occur
Unknown, reported today in-game, stated to have been going on 'forever'.
#### Client version, Server revision & Game ID
Client Version: 513
Server Revision: 02aeb12d1f76e63b403f40b988cd8450ba17ec5a - dev - 2021-07-21
Game ID: cda-ahmg
Current map: SEV Torch
#### Issue bingo
<!-- Check these by writing an x inside the [ ] (like this: [x])-->
<!-- Don't forget to remove the space between the brackets, or it won't work! -->
- [X] Issue could be reproduced at least once
- [X] Issue could be reproduced by different players
- [X] Issue could be reproduced in multiple rounds
- [X] Issue happened in a recent (less than 7 days ago) round
- [X] [Couldn't find an existing issue about this](https://github.com/Baystation12/Baystation12/issues)
| non_infrastructure | mapped cargo tug self links incorrectly description of issue the cargo tug and train mapped into the cargo warehouse do not link themselves correctly when initialized the central trolley hitches itself both to the cargo engine and the third trolley while the engine and trailing trolley both hitch themselves to the central trolley resulting in a tangled mess of a b c when the code expected a b c this in turn causes the tug to not move and instead eject a user when they try to drive it difference between expected and actual behavior engine and trolleys should correctly chain themselves instead of creating broken cross links workaround use the right click menu to unhitch the engine and both trolleys then reconnect everything together with click drag this will allow the tug to be driveable again steps to reproduce start the torch map go to the cargo warehouse try to use the cargo tug specific information for locating engine obj vehicle train cargo engine trolley obj vehicle train cargo trolley very likely an issue with the base obj vehicle train type length of time in which bug has been known to occur unknown reported today in game stated to have been going on forever client version server revision game id client version server revision dev game id cda ahmg current map sev torch issue bingo issue could be reproduced at least once issue could be reproduced by different players issue could be reproduced in multiple rounds issue happened in a recent less than days ago round | 0 |
2,509 | 3,734,068,853 | IssuesEvent | 2016-03-08 04:04:08 | dotnet/roslyn | https://api.github.com/repos/dotnet/roslyn | closed | [Test Failure] Individual TAO Test Failure in prtest/win/vsi/p2 on PR #651 | Area-Infrastructure Bug Contributor Pain | PR: [#651](https://github.com/dotnet/roslyn-internal/pull/651) *Merge stabilization into future-stabilization* by @vslsnap
Failure: http://dotnet-ci.cloudapp.net/job/roslyn-internal_prtest_win_vsi_p2/174/
**1 Test Failures:**
CSharpGotoDefinition.Go To Def opens in a provisional tab if document wasn't already open
**Issue 1: Individual TAO Test Failure**
Failing integration tests:
CSharpGotoDefinition.xml | 1.0 | [Test Failure] Individual TAO Test Failure in prtest/win/vsi/p2 on PR #651 - PR: [#651](https://github.com/dotnet/roslyn-internal/pull/651) *Merge stabilization into future-stabilization* by @vslsnap
Failure: http://dotnet-ci.cloudapp.net/job/roslyn-internal_prtest_win_vsi_p2/174/
**1 Test Failures:**
CSharpGotoDefinition.Go To Def opens in a provisional tab if document wasn't already open
**Issue 1: Individual TAO Test Failure**
Failing integration tests:
CSharpGotoDefinition.xml | infrastructure | individual tao test failure in prtest win vsi on pr pr merge stabilization into future stabilization by vslsnap failure test failures csharpgotodefinition go to def opens in a provisional tab if document wasn t already open issue individual tao test failure failing integration tests csharpgotodefinition xml | 1 |
22,293 | 15,088,109,954 | IssuesEvent | 2021-02-05 23:46:53 | near/multisig-tool | https://api.github.com/repos/near/multisig-tool | opened | [DEV] Setup architecture | Infrastructure | @eclipseeer please write here the details of how it's gonna work after you set it up | 1.0 | [DEV] Setup architecture - @eclipseeer please write here the details of how it's gonna work after you set it up | infrastructure | setup architecture eclipseeer please write here the details of how it s gonna work after you set it up | 1 |
27,117 | 21,186,110,377 | IssuesEvent | 2022-04-08 12:55:26 | alkem-io/alkemio | https://api.github.com/repos/alkem-io/alkemio | closed | ELK stack PoC | Infrastructure | ## Description
As a DevOps engineer
I want to have aggregated logs
so that I can monitor my services and not individual pods and I can have an overview of the whole application.
## Acceptance criteria
- [x] Elk stack enabled in Dev environmet
- [x] Log aggregation enabled
- [x] Kibana dashboard available accessible from internet
- [ ] File logs - e.g. subscriptions logContext - can be visualized from Kibana
- [x] Document what do we want to achieve and how is it achieved in this PoC in an MD file
## Additional Context
Timebox to 1.5 days
| 1.0 | ELK stack PoC - ## Description
As a DevOps engineer
I want to have aggregated logs
so that I can monitor my services and not individual pods and I can have an overview of the whole application.
## Acceptance criteria
- [x] Elk stack enabled in Dev environmet
- [x] Log aggregation enabled
- [x] Kibana dashboard available accessible from internet
- [ ] File logs - e.g. subscriptions logContext - can be visualized from Kibana
- [x] Document what do we want to achieve and how is it achieved in this PoC in an MD file
## Additional Context
Timebox to 1.5 days
| infrastructure | elk stack poc description as a devops engineer i want to have aggregated logs so that i can monitor my services and not individual pods and i can have an overview of the whole application acceptance criteria elk stack enabled in dev environmet log aggregation enabled kibana dashboard available accessible from internet file logs e g subscriptions logcontext can be visualized from kibana document what do we want to achieve and how is it achieved in this poc in an md file additional context timebox to days | 1 |
36,508 | 7,974,042,250 | IssuesEvent | 2018-07-17 02:54:13 | Microsoft/devkit-sdk | https://api.github.com/repos/Microsoft/devkit-sdk | closed | 1st screen is not friendly | Bug bash - 1.1.0 Defect P2 | Currently when upgrade to 1.1 firmware, it's super confusion to returning users - where is my sensor status app?
The hint shows on the screen makes me feel I did something wrong.
Suggestion - "Connect WiFi to activate your DevKit"
| 1.0 | 1st screen is not friendly - Currently when upgrade to 1.1 firmware, it's super confusion to returning users - where is my sensor status app?
The hint shows on the screen makes me feel I did something wrong.
Suggestion - "Connect WiFi to activate your DevKit"
| non_infrastructure | screen is not friendly currently when upgrade to firmware it s super confusion to returning users where is my sensor status app the hint shows on the screen makes me feel i did something wrong suggestion connect wifi to activate your devkit | 0 |
266,049 | 28,298,880,079 | IssuesEvent | 2023-04-10 02:51:09 | nidhi7598/linux-4.19.72 | https://api.github.com/repos/nidhi7598/linux-4.19.72 | closed | CVE-2019-19073 (Medium) detected in linuxlinux-4.19.254 - autoclosed | Mend: dependency security vulnerability | ## CVE-2019-19073 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.254</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.19.72/commit/10a8c99e4f60044163c159867bc6f5452c1c36e5">10a8c99e4f60044163c159867bc6f5452c1c36e5</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/wireless/ath/ath9k/htc_hst.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/wireless/ath/ath9k/htc_hst.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Memory leaks in drivers/net/wireless/ath/ath9k/htc_hst.c in the Linux kernel through 5.3.11 allow attackers to cause a denial of service (memory consumption) by triggering wait_for_completion_timeout() failures. This affects the htc_config_pipe_credits() function, the htc_setup_complete() function, and the htc_connect_service() function, aka CID-853acf7caf10.
<p>Publish Date: 2019-11-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-19073>CVE-2019-19073</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19073">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19073</a></p>
<p>Release Date: 2019-11-18</p>
<p>Fix Resolution: v5.4-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-19073 (Medium) detected in linuxlinux-4.19.254 - autoclosed - ## CVE-2019-19073 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.254</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.19.72/commit/10a8c99e4f60044163c159867bc6f5452c1c36e5">10a8c99e4f60044163c159867bc6f5452c1c36e5</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/wireless/ath/ath9k/htc_hst.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/wireless/ath/ath9k/htc_hst.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Memory leaks in drivers/net/wireless/ath/ath9k/htc_hst.c in the Linux kernel through 5.3.11 allow attackers to cause a denial of service (memory consumption) by triggering wait_for_completion_timeout() failures. This affects the htc_config_pipe_credits() function, the htc_setup_complete() function, and the htc_connect_service() function, aka CID-853acf7caf10.
<p>Publish Date: 2019-11-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-19073>CVE-2019-19073</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19073">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19073</a></p>
<p>Release Date: 2019-11-18</p>
<p>Fix Resolution: v5.4-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | cve medium detected in linuxlinux autoclosed cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files drivers net wireless ath htc hst c drivers net wireless ath htc hst c vulnerability details memory leaks in drivers net wireless ath htc hst c in the linux kernel through allow attackers to cause a denial of service memory consumption by triggering wait for completion timeout failures this affects the htc config pipe credits function the htc setup complete function and the htc connect service function aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
447,052 | 31,593,375,413 | IssuesEvent | 2023-09-05 02:01:30 | wixtoolset/issues | https://api.github.com/repos/wixtoolset/issues | closed | `Package/@Scope` XSD doesn't allow for preprocessor variables | documentation | If this issue is a bug:
* Which version of WiX are you building with?
> v4.0.1
* Which version of Visual Studio are you building with (if any)?
> 17.6.5
* Which version of the WiX Toolset Visual Studio Extension are you building with (if any)?
> HeatWave v1.0.1
* Which version of .NET are you building with?
> n/a
* If the problem occurs when installing your packages built with WiX, what is the version of Windows the package is running on?
> n/a
* Describe the problem and the steps to reproduce it.
> wix.xsd for `Package/@Scope` is declared as an enum so Visual Studio complains about preprocessor variables.
* Describe the behavior you expected and how it differed from the actual behavior.
> There should not be warnings in the VS Error List. Needs something like YesNoTypeUnion.
| 1.0 | `Package/@Scope` XSD doesn't allow for preprocessor variables - If this issue is a bug:
* Which version of WiX are you building with?
> v4.0.1
* Which version of Visual Studio are you building with (if any)?
> 17.6.5
* Which version of the WiX Toolset Visual Studio Extension are you building with (if any)?
> HeatWave v1.0.1
* Which version of .NET are you building with?
> n/a
* If the problem occurs when installing your packages built with WiX, what is the version of Windows the package is running on?
> n/a
* Describe the problem and the steps to reproduce it.
> wix.xsd for `Package/@Scope` is declared as an enum so Visual Studio complains about preprocessor variables.
* Describe the behavior you expected and how it differed from the actual behavior.
> There should not be warnings in the VS Error List. Needs something like YesNoTypeUnion.
| non_infrastructure | package scope xsd doesn t allow for preprocessor variables if this issue is a bug which version of wix are you building with which version of visual studio are you building with if any which version of the wix toolset visual studio extension are you building with if any heatwave which version of net are you building with n a if the problem occurs when installing your packages built with wix what is the version of windows the package is running on n a describe the problem and the steps to reproduce it wix xsd for package scope is declared as an enum so visual studio complains about preprocessor variables describe the behavior you expected and how it differed from the actual behavior there should not be warnings in the vs error list needs something like yesnotypeunion | 0 |
51,202 | 6,150,845,847 | IssuesEvent | 2017-06-28 00:00:23 | mozilla/activity-stream | https://api.github.com/repos/mozilla/activity-stream | closed | Dashboard for Activity Stream Shield vs. Activity Stream Test Pilot | P2 Test Pilot | This is to give us an idea of whether release users behave similar to test pilot users. | 1.0 | Dashboard for Activity Stream Shield vs. Activity Stream Test Pilot - This is to give us an idea of whether release users behave similar to test pilot users. | non_infrastructure | dashboard for activity stream shield vs activity stream test pilot this is to give us an idea of whether release users behave similar to test pilot users | 0 |
2,109 | 3,512,500,071 | IssuesEvent | 2016-01-11 01:36:14 | MultiMC/MultiMC5 | https://api.github.com/repos/MultiMC/MultiMC5 | opened | Check help pages for correctness | infrastructure | * Make sure every dialog page has the corresponding help page
* Make sure the help pages do not contain nonsense or outdated information | 1.0 | Check help pages for correctness - * Make sure every dialog page has the corresponding help page
* Make sure the help pages do not contain nonsense or outdated information | infrastructure | check help pages for correctness make sure every dialog page has the corresponding help page make sure the help pages do not contain nonsense or outdated information | 1 |
194,589 | 22,262,039,895 | IssuesEvent | 2022-06-10 02:01:25 | Trinadh465/device_renesas_kernel_AOSP10_r33_CVE-2021-33034 | https://api.github.com/repos/Trinadh465/device_renesas_kernel_AOSP10_r33_CVE-2021-33034 | reopened | CVE-2019-19065 (Medium) detected in linuxlinux-4.19.239 | security vulnerability | ## CVE-2019-19065 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.239</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Trinadh465/device_renesas_kernel_AOSP10_r33_CVE-2021-33034/commit/19525e8c58fe9ba0d7cb0f7a1a87d31d30380de6">19525e8c58fe9ba0d7cb0f7a1a87d31d30380de6</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/infiniband/hw/hfi1/sdma.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/infiniband/hw/hfi1/sdma.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
** DISPUTED ** A memory leak in the sdma_init() function in drivers/infiniband/hw/hfi1/sdma.c in the Linux kernel before 5.3.9 allows attackers to cause a denial of service (memory consumption) by triggering rhashtable_init() failures, aka CID-34b3be18a04e. NOTE: This has been disputed as not a vulnerability because "rhashtable_init() can only fail if it is passed invalid values in the second parameter's struct, but when invoked from sdma_init() that is a pointer to a static const struct, so an attacker could only trigger failure if they could corrupt kernel memory (in which case a small memory leak is not a significant problem)."
<p>Publish Date: 2019-11-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19065>CVE-2019-19065</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19065">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19065</a></p>
<p>Release Date: 2020-08-24</p>
<p>Fix Resolution: v5.4-rc3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-19065 (Medium) detected in linuxlinux-4.19.239 - ## CVE-2019-19065 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.239</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Trinadh465/device_renesas_kernel_AOSP10_r33_CVE-2021-33034/commit/19525e8c58fe9ba0d7cb0f7a1a87d31d30380de6">19525e8c58fe9ba0d7cb0f7a1a87d31d30380de6</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/infiniband/hw/hfi1/sdma.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/infiniband/hw/hfi1/sdma.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
** DISPUTED ** A memory leak in the sdma_init() function in drivers/infiniband/hw/hfi1/sdma.c in the Linux kernel before 5.3.9 allows attackers to cause a denial of service (memory consumption) by triggering rhashtable_init() failures, aka CID-34b3be18a04e. NOTE: This has been disputed as not a vulnerability because "rhashtable_init() can only fail if it is passed invalid values in the second parameter's struct, but when invoked from sdma_init() that is a pointer to a static const struct, so an attacker could only trigger failure if they could corrupt kernel memory (in which case a small memory leak is not a significant problem)."
<p>Publish Date: 2019-11-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19065>CVE-2019-19065</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19065">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19065</a></p>
<p>Release Date: 2020-08-24</p>
<p>Fix Resolution: v5.4-rc3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_infrastructure | cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files drivers infiniband hw sdma c drivers infiniband hw sdma c vulnerability details disputed a memory leak in the sdma init function in drivers infiniband hw sdma c in the linux kernel before allows attackers to cause a denial of service memory consumption by triggering rhashtable init failures aka cid note this has been disputed as not a vulnerability because rhashtable init can only fail if it is passed invalid values in the second parameter s struct but when invoked from sdma init that is a pointer to a static const struct so an attacker could only trigger failure if they could corrupt kernel memory in which case a small memory leak is not a significant problem publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
25,968 | 19,520,842,552 | IssuesEvent | 2021-12-29 18:08:36 | OHNLP/N3C-NLP-Documentation | https://api.github.com/repos/OHNLP/N3C-NLP-Documentation | closed | Context rules for template languages | Infrastructure Algorithm | - [ ] Compose a few context rules that works for template languages
- [ ] ": Not present"
- [ ] Test | 1.0 | Context rules for template languages - - [ ] Compose a few context rules that works for template languages
- [ ] ": Not present"
- [ ] Test | infrastructure | context rules for template languages compose a few context rules that works for template languages not present test | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.