Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
757
labels
stringlengths
4
664
body
stringlengths
3
261k
index
stringclasses
10 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
232k
binary_label
int64
0
1
264,368
23,114,725,588
IssuesEvent
2022-07-27 15:40:42
Lightning-AI/lightning
https://api.github.com/repos/Lightning-AI/lightning
closed
The TPU issues in Lightning
ci accelerator: tpu tests
## 🐛 Bug Recent observations have made it clear that there are many problems with either the TPU implementation in Lightning or the test environment: 1. Not all tests written in Lightning for TPU are executed. Only a hand-maintained list of tests ever runs. #11098 2. Attempting to address 1) reveals further that among the tests that do run, there are many decorated with a wrapper `@pl_multi_process_test`, which suppresses assertion errors and exceptions of broken tests. The result is that we have a lot of tests that are broken but never surface in the CI. ### To Reproduce A simple way to reproduce this is by removing all decorators, which is what I have done in #11098, and then let the tests run and fail. Attached is the full log file of such a CI run: [tpu-logs-without-pl-multi.txt](https://github.com/Lightning-AI/lightning/files/9120671/tpu-logs-without-pl-multi.txt) In summary: 17 failed, 48 passed ``` FAILED tests/tests_pytorch/callbacks/test_device_stats_monitor.py::test_device_stats_monitor_tpu FAILED tests/tests_pytorch/models/test_tpu.py::test_model_tpu_index[1] - Runt... FAILED tests/tests_pytorch/models/test_tpu.py::test_model_tpu_index[5] - Runt... FAILED tests/tests_pytorch/models/test_tpu.py::test_model_tpu_devices_8 - tor... FAILED tests/tests_pytorch/models/test_tpu.py::test_model_16bit_tpu_index[1] FAILED tests/tests_pytorch/models/test_tpu.py::test_model_16bit_tpu_index[5] FAILED tests/tests_pytorch/models/test_tpu.py::test_model_16bit_tpu_devices_8 FAILED tests/tests_pytorch/models/test_tpu.py::test_model_tpu_early_stop - to... FAILED tests/tests_pytorch/models/test_tpu.py::test_dataloaders_passed_to_fit FAILED tests/tests_pytorch/models/test_tpu.py::test_broadcast_on_tpu - torch.... FAILED tests/tests_pytorch/models/test_tpu.py::test_tpu_reduce - torch.multip... FAILED tests/tests_pytorch/models/test_tpu.py::test_if_test_works_with_checkpoint_false FAILED tests/tests_pytorch/models/test_tpu.py::test_tpu_sync_dist - torch.mul... FAILED tests/tests_pytorch/models/test_tpu.py::test_tpu_debug_mode - torch.mu... FAILED tests/tests_pytorch/models/test_tpu.py::test_tpu_host_world_size - tor... FAILED tests/tests_pytorch/profilers/test_xla_profiler.py::test_xla_profiler_instance FAILED tests/tests_pytorch/trainer/properties/test_estimated_stepping_batches.py::test_num_stepping_batches_with_tpu[8-8] ERROR tests/tests_pytorch/models/test_tpu.py::test_model_16bit_tpu_index[1] ERROR tests/tests_pytorch/models/test_tpu.py::test_model_16bit_tpu_index[5] ``` There is of course the infamous cryptic error message for several test cases `Exception in device=TPU:2: Cannot replicate if number of devices (1) is different from 8` Which sometimes hints at the possibility that we are accessing the `xm.xla_device` before spawning processes. Other examples: 1) ``` self = <pytorch_lightning.trainer.connectors.accelerator_connector.AcceleratorConnector object at 0x7f7bbead03d0> @property def is_distributed(self) -> bool: # TODO: deprecate this property # Used for custom plugins. # Custom plugins should implement is_distributed property. if hasattr(self.strategy, "is_distributed") and not isinstance(self.accelerator, TPUAccelerator): return self.strategy.is_distributed distributed_strategy = ( DDP2Strategy, DDPStrategy, DDPSpawnShardedStrategy, DDPShardedStrategy, DDPFullyShardedNativeStrategy, DDPFullyShardedStrategy, DDPSpawnStrategy, DeepSpeedStrategy, TPUSpawnStrategy, HorovodStrategy, HPUParallelStrategy, ) is_distributed = isinstance(self.strategy, distributed_strategy) if isinstance(self.accelerator, TPUAccelerator): > is_distributed |= self.strategy.is_distributed E TypeError: unsupported operand type(s) for |=: 'bool' and 'NoneType' ``` 2) ``` def has_len_all_ranks( dataloader: DataLoader, training_type: "pl.Strategy", model: Union["pl.LightningModule", "pl.LightningDataModule"], ) -> bool: """Checks if a given Dataloader has ``__len__`` method implemented i.e. if it is a finite dataloader or infinite dataloader.""" try: local_length = len(dataloader) total_length = training_type.reduce(torch.tensor(local_length).to(model.device), reduce_op="sum") > if total_length == 0: E RuntimeError: Not found: From /job:tpu_worker/replica:0/task:0: E 2 root error(s) found. E (0) Not found: No subgraph found for uid 2894109085761937038 E [[{{node XRTExecute}}]] E (1) Not found: No subgraph found for uid 2894109085761937038 E [[{{node XRTExecute}}]] E [[XRTExecute_G29]] E 0 successful operations. E 0 derived errors ignored. ``` Furthermore, sometimes, non-deterministically, the CI just stops in the middle of execution: ``` .... profilers/test_xla_profiler.py::test_xla_profiler_instance FAILED [ 93%] strategies/test_tpu_spawn.py::test_model_tpu_one_core PASSED [ 95%] Done with log retrieval attempt. Exited with code exit status 2 CircleCI received exit code 2 ``` ### Expected behavior It is unclear what the intention was when designing the test setup. The decorators were introduced way back in #2512 and have never much changed. Meanwhile, strategy and accelerators have undergone major design changes and countless refactors. I propose to re-evaluate whether the `pl_multi_process_test` decorator is still needed, and if so, document why, how to use it and when to use it correctly. ### Possible Action My suggestion is to 1. Remove the decorator 2. Debug each test on the VM 3. Run tests that require it in standalone mode 4. Reduce the verbosity of the mind boggling thousands of nonsense lines printed in the CI 5. Upgrade to the [latest XLA and PyTorch version](https://github.com/pytorch/xla/releases/tag/v1.12.0) cc @carmocca @akihironitta @borda @kaushikb11 @rohitgr7
1.0
The TPU issues in Lightning - ## 🐛 Bug Recent observations have made it clear that there are many problems with either the TPU implementation in Lightning or the test environment: 1. Not all tests written in Lightning for TPU are executed. Only a hand-maintained list of tests ever runs. #11098 2. Attempting to address 1) reveals further that among the tests that do run, there are many decorated with a wrapper `@pl_multi_process_test`, which suppresses assertion errors and exceptions of broken tests. The result is that we have a lot of tests that are broken but never surface in the CI. ### To Reproduce A simple way to reproduce this is by removing all decorators, which is what I have done in #11098, and then let the tests run and fail. Attached is the full log file of such a CI run: [tpu-logs-without-pl-multi.txt](https://github.com/Lightning-AI/lightning/files/9120671/tpu-logs-without-pl-multi.txt) In summary: 17 failed, 48 passed ``` FAILED tests/tests_pytorch/callbacks/test_device_stats_monitor.py::test_device_stats_monitor_tpu FAILED tests/tests_pytorch/models/test_tpu.py::test_model_tpu_index[1] - Runt... FAILED tests/tests_pytorch/models/test_tpu.py::test_model_tpu_index[5] - Runt... FAILED tests/tests_pytorch/models/test_tpu.py::test_model_tpu_devices_8 - tor... FAILED tests/tests_pytorch/models/test_tpu.py::test_model_16bit_tpu_index[1] FAILED tests/tests_pytorch/models/test_tpu.py::test_model_16bit_tpu_index[5] FAILED tests/tests_pytorch/models/test_tpu.py::test_model_16bit_tpu_devices_8 FAILED tests/tests_pytorch/models/test_tpu.py::test_model_tpu_early_stop - to... FAILED tests/tests_pytorch/models/test_tpu.py::test_dataloaders_passed_to_fit FAILED tests/tests_pytorch/models/test_tpu.py::test_broadcast_on_tpu - torch.... FAILED tests/tests_pytorch/models/test_tpu.py::test_tpu_reduce - torch.multip... FAILED tests/tests_pytorch/models/test_tpu.py::test_if_test_works_with_checkpoint_false FAILED tests/tests_pytorch/models/test_tpu.py::test_tpu_sync_dist - torch.mul... FAILED tests/tests_pytorch/models/test_tpu.py::test_tpu_debug_mode - torch.mu... FAILED tests/tests_pytorch/models/test_tpu.py::test_tpu_host_world_size - tor... FAILED tests/tests_pytorch/profilers/test_xla_profiler.py::test_xla_profiler_instance FAILED tests/tests_pytorch/trainer/properties/test_estimated_stepping_batches.py::test_num_stepping_batches_with_tpu[8-8] ERROR tests/tests_pytorch/models/test_tpu.py::test_model_16bit_tpu_index[1] ERROR tests/tests_pytorch/models/test_tpu.py::test_model_16bit_tpu_index[5] ``` There is of course the infamous cryptic error message for several test cases `Exception in device=TPU:2: Cannot replicate if number of devices (1) is different from 8` Which sometimes hints at the possibility that we are accessing the `xm.xla_device` before spawning processes. Other examples: 1) ``` self = <pytorch_lightning.trainer.connectors.accelerator_connector.AcceleratorConnector object at 0x7f7bbead03d0> @property def is_distributed(self) -> bool: # TODO: deprecate this property # Used for custom plugins. # Custom plugins should implement is_distributed property. if hasattr(self.strategy, "is_distributed") and not isinstance(self.accelerator, TPUAccelerator): return self.strategy.is_distributed distributed_strategy = ( DDP2Strategy, DDPStrategy, DDPSpawnShardedStrategy, DDPShardedStrategy, DDPFullyShardedNativeStrategy, DDPFullyShardedStrategy, DDPSpawnStrategy, DeepSpeedStrategy, TPUSpawnStrategy, HorovodStrategy, HPUParallelStrategy, ) is_distributed = isinstance(self.strategy, distributed_strategy) if isinstance(self.accelerator, TPUAccelerator): > is_distributed |= self.strategy.is_distributed E TypeError: unsupported operand type(s) for |=: 'bool' and 'NoneType' ``` 2) ``` def has_len_all_ranks( dataloader: DataLoader, training_type: "pl.Strategy", model: Union["pl.LightningModule", "pl.LightningDataModule"], ) -> bool: """Checks if a given Dataloader has ``__len__`` method implemented i.e. if it is a finite dataloader or infinite dataloader.""" try: local_length = len(dataloader) total_length = training_type.reduce(torch.tensor(local_length).to(model.device), reduce_op="sum") > if total_length == 0: E RuntimeError: Not found: From /job:tpu_worker/replica:0/task:0: E 2 root error(s) found. E (0) Not found: No subgraph found for uid 2894109085761937038 E [[{{node XRTExecute}}]] E (1) Not found: No subgraph found for uid 2894109085761937038 E [[{{node XRTExecute}}]] E [[XRTExecute_G29]] E 0 successful operations. E 0 derived errors ignored. ``` Furthermore, sometimes, non-deterministically, the CI just stops in the middle of execution: ``` .... profilers/test_xla_profiler.py::test_xla_profiler_instance FAILED [ 93%] strategies/test_tpu_spawn.py::test_model_tpu_one_core PASSED [ 95%] Done with log retrieval attempt. Exited with code exit status 2 CircleCI received exit code 2 ``` ### Expected behavior It is unclear what the intention was when designing the test setup. The decorators were introduced way back in #2512 and have never much changed. Meanwhile, strategy and accelerators have undergone major design changes and countless refactors. I propose to re-evaluate whether the `pl_multi_process_test` decorator is still needed, and if so, document why, how to use it and when to use it correctly. ### Possible Action My suggestion is to 1. Remove the decorator 2. Debug each test on the VM 3. Run tests that require it in standalone mode 4. Reduce the verbosity of the mind boggling thousands of nonsense lines printed in the CI 5. Upgrade to the [latest XLA and PyTorch version](https://github.com/pytorch/xla/releases/tag/v1.12.0) cc @carmocca @akihironitta @borda @kaushikb11 @rohitgr7
non_defect
the tpu issues in lightning 🐛 bug recent observations have made it clear that there are many problems with either the tpu implementation in lightning or the test environment not all tests written in lightning for tpu are executed only a hand maintained list of tests ever runs attempting to address reveals further that among the tests that do run there are many decorated with a wrapper pl multi process test which suppresses assertion errors and exceptions of broken tests the result is that we have a lot of tests that are broken but never surface in the ci to reproduce a simple way to reproduce this is by removing all decorators which is what i have done in and then let the tests run and fail attached is the full log file of such a ci run in summary failed passed failed tests tests pytorch callbacks test device stats monitor py test device stats monitor tpu failed tests tests pytorch models test tpu py test model tpu index runt failed tests tests pytorch models test tpu py test model tpu index runt failed tests tests pytorch models test tpu py test model tpu devices tor failed tests tests pytorch models test tpu py test model tpu index failed tests tests pytorch models test tpu py test model tpu index failed tests tests pytorch models test tpu py test model tpu devices failed tests tests pytorch models test tpu py test model tpu early stop to failed tests tests pytorch models test tpu py test dataloaders passed to fit failed tests tests pytorch models test tpu py test broadcast on tpu torch failed tests tests pytorch models test tpu py test tpu reduce torch multip failed tests tests pytorch models test tpu py test if test works with checkpoint false failed tests tests pytorch models test tpu py test tpu sync dist torch mul failed tests tests pytorch models test tpu py test tpu debug mode torch mu failed tests tests pytorch models test tpu py test tpu host world size tor failed tests tests pytorch profilers test xla profiler py test xla profiler instance failed tests tests pytorch trainer properties test estimated stepping batches py test num stepping batches with tpu error tests tests pytorch models test tpu py test model tpu index error tests tests pytorch models test tpu py test model tpu index there is of course the infamous cryptic error message for several test cases exception in device tpu cannot replicate if number of devices is different from which sometimes hints at the possibility that we are accessing the xm xla device before spawning processes other examples self property def is distributed self bool todo deprecate this property used for custom plugins custom plugins should implement is distributed property if hasattr self strategy is distributed and not isinstance self accelerator tpuaccelerator return self strategy is distributed distributed strategy ddpstrategy ddpspawnshardedstrategy ddpshardedstrategy ddpfullyshardednativestrategy ddpfullyshardedstrategy ddpspawnstrategy deepspeedstrategy tpuspawnstrategy horovodstrategy hpuparallelstrategy is distributed isinstance self strategy distributed strategy if isinstance self accelerator tpuaccelerator is distributed self strategy is distributed e typeerror unsupported operand type s for bool and nonetype def has len all ranks dataloader dataloader training type pl strategy model union bool checks if a given dataloader has len method implemented i e if it is a finite dataloader or infinite dataloader try local length len dataloader total length training type reduce torch tensor local length to model device reduce op sum if total length e runtimeerror not found from job tpu worker replica task e root error s found e not found no subgraph found for uid e e not found no subgraph found for uid e e e successful operations e derived errors ignored furthermore sometimes non deterministically the ci just stops in the middle of execution profilers test xla profiler py test xla profiler instance failed strategies test tpu spawn py test model tpu one core passed done with log retrieval attempt exited with code exit status circleci received exit code expected behavior it is unclear what the intention was when designing the test setup the decorators were introduced way back in and have never much changed meanwhile strategy and accelerators have undergone major design changes and countless refactors i propose to re evaluate whether the pl multi process test decorator is still needed and if so document why how to use it and when to use it correctly possible action my suggestion is to remove the decorator debug each test on the vm run tests that require it in standalone mode reduce the verbosity of the mind boggling thousands of nonsense lines printed in the ci upgrade to the cc carmocca akihironitta borda
0
35,636
7,794,821,824
IssuesEvent
2018-06-08 05:14:17
StrikeNP/trac_test
https://api.github.com/repos/StrikeNP/trac_test
closed
Capitalize "Unreleased" in src/Benchmark_cases/unreleased_cases (Trac #84)
Migrated from Trac clubb_src defect senkbeil@uwm.edu
For consistency, we should capitalize the names of all subdirectories of src. The only change needed is for subdirectory unreleased_cases. This will require modifications to the code browser and perhaps other things related to the nightly tests. Attachments: Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/84 ```json { "status": "closed", "changetime": "2009-06-23T14:22:19", "description": "For consistency, we should capitalize the names of all subdirectories of src. The only change needed is for subdirectory unreleased_cases.\n\nThis will require modifications to the code browser and perhaps other things related to the nightly tests.", "reporter": "vlarson@uwm.edu", "cc": "", "resolution": "Verified by V. Larson", "_ts": "1245766939000000", "component": "clubb_src", "summary": "Capitalize \"Unreleased\" in src/Benchmark_cases/unreleased_cases", "priority": "minor", "keywords": "", "time": "2009-06-19T20:58:35", "milestone": "3. Refactor CLUBB", "owner": "senkbeil@uwm.edu", "type": "defect" } ```
1.0
Capitalize "Unreleased" in src/Benchmark_cases/unreleased_cases (Trac #84) - For consistency, we should capitalize the names of all subdirectories of src. The only change needed is for subdirectory unreleased_cases. This will require modifications to the code browser and perhaps other things related to the nightly tests. Attachments: Migrated from http://carson.math.uwm.edu/trac/clubb/ticket/84 ```json { "status": "closed", "changetime": "2009-06-23T14:22:19", "description": "For consistency, we should capitalize the names of all subdirectories of src. The only change needed is for subdirectory unreleased_cases.\n\nThis will require modifications to the code browser and perhaps other things related to the nightly tests.", "reporter": "vlarson@uwm.edu", "cc": "", "resolution": "Verified by V. Larson", "_ts": "1245766939000000", "component": "clubb_src", "summary": "Capitalize \"Unreleased\" in src/Benchmark_cases/unreleased_cases", "priority": "minor", "keywords": "", "time": "2009-06-19T20:58:35", "milestone": "3. Refactor CLUBB", "owner": "senkbeil@uwm.edu", "type": "defect" } ```
defect
capitalize unreleased in src benchmark cases unreleased cases trac for consistency we should capitalize the names of all subdirectories of src the only change needed is for subdirectory unreleased cases this will require modifications to the code browser and perhaps other things related to the nightly tests attachments migrated from json status closed changetime description for consistency we should capitalize the names of all subdirectories of src the only change needed is for subdirectory unreleased cases n nthis will require modifications to the code browser and perhaps other things related to the nightly tests reporter vlarson uwm edu cc resolution verified by v larson ts component clubb src summary capitalize unreleased in src benchmark cases unreleased cases priority minor keywords time milestone refactor clubb owner senkbeil uwm edu type defect
1
132,386
10,743,083,940
IssuesEvent
2019-10-30 00:45:53
elastic/kibana
https://api.github.com/repos/elastic/kibana
opened
Failing test: Firefox XPack UI Functional Tests.x-pack/test/functional/apps/machine_learning/anomaly_detection/saved_search_job·ts - machine learning anomaly detection saved search with lucene query job creation displays the created job in the job list
failed-test
A test failed on a tracked branch ``` Error: retry.tryForTime timeout: Error: retry.try timeout: TimeoutError: Waiting for element to be located By(css selector, [data-test-subj="mlMainTab anomalyDetection"]) Wait timed out after 10007ms at /dev/shm/workspace/kibana/node_modules/selenium-webdriver/lib/webdriver.js:841:17 at process._tickCallback (internal/process/next_tick.js:68:7) at lastError (/dev/shm/workspace/kibana/test/common/services/retry/retry_for_success.ts:28:9) at onFailure (/dev/shm/workspace/kibana/test/common/services/retry/retry_for_success.ts:68:13) at lastError (/dev/shm/workspace/kibana/test/common/services/retry/retry_for_success.ts:28:9) at onFailure (/dev/shm/workspace/kibana/test/common/services/retry/retry_for_success.ts:68:13) ``` First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+7.x/845/) <!-- kibanaCiData = {"failed-test":{"test.class":"Firefox XPack UI Functional Tests.x-pack/test/functional/apps/machine_learning/anomaly_detection/saved_search_job·ts","test.name":"machine learning anomaly detection saved search with lucene query job creation displays the created job in the job list","test.failCount":1}} -->
1.0
Failing test: Firefox XPack UI Functional Tests.x-pack/test/functional/apps/machine_learning/anomaly_detection/saved_search_job·ts - machine learning anomaly detection saved search with lucene query job creation displays the created job in the job list - A test failed on a tracked branch ``` Error: retry.tryForTime timeout: Error: retry.try timeout: TimeoutError: Waiting for element to be located By(css selector, [data-test-subj="mlMainTab anomalyDetection"]) Wait timed out after 10007ms at /dev/shm/workspace/kibana/node_modules/selenium-webdriver/lib/webdriver.js:841:17 at process._tickCallback (internal/process/next_tick.js:68:7) at lastError (/dev/shm/workspace/kibana/test/common/services/retry/retry_for_success.ts:28:9) at onFailure (/dev/shm/workspace/kibana/test/common/services/retry/retry_for_success.ts:68:13) at lastError (/dev/shm/workspace/kibana/test/common/services/retry/retry_for_success.ts:28:9) at onFailure (/dev/shm/workspace/kibana/test/common/services/retry/retry_for_success.ts:68:13) ``` First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+7.x/845/) <!-- kibanaCiData = {"failed-test":{"test.class":"Firefox XPack UI Functional Tests.x-pack/test/functional/apps/machine_learning/anomaly_detection/saved_search_job·ts","test.name":"machine learning anomaly detection saved search with lucene query job creation displays the created job in the job list","test.failCount":1}} -->
non_defect
failing test firefox xpack ui functional tests x pack test functional apps machine learning anomaly detection saved search job·ts machine learning anomaly detection saved search with lucene query job creation displays the created job in the job list a test failed on a tracked branch error retry tryfortime timeout error retry try timeout timeouterror waiting for element to be located by css selector wait timed out after at dev shm workspace kibana node modules selenium webdriver lib webdriver js at process tickcallback internal process next tick js at lasterror dev shm workspace kibana test common services retry retry for success ts at onfailure dev shm workspace kibana test common services retry retry for success ts at lasterror dev shm workspace kibana test common services retry retry for success ts at onfailure dev shm workspace kibana test common services retry retry for success ts first failure
0
379,329
26,366,460,067
IssuesEvent
2023-01-11 16:56:51
Real-Dev-Squad/website-api-contracts
https://api.github.com/repos/Real-Dev-Squad/website-api-contracts
closed
API contract for log APIs
documentation
## What needs to be done? - Create API contract for `/cache` - Create API contract for `/logs/:type`
1.0
API contract for log APIs - ## What needs to be done? - Create API contract for `/cache` - Create API contract for `/logs/:type`
non_defect
api contract for log apis what needs to be done create api contract for cache create api contract for logs type
0
65,891
19,766,751,021
IssuesEvent
2022-01-17 04:04:03
vector-im/element-android
https://api.github.com/repos/vector-im/element-android
opened
Glitch when sending Latex
T-Defect
### Steps to reproduce 1. Enable Latex in Labs and enable markdown 2. Send any latex formatted code 3. The sceeen stutters and glitches while displaying the latex. Video: https://user-images.githubusercontent.com/62639087/149706093-b0a42bf8-cb15-4c50-a8f4-fd5334eb9ec1.mp4 Please ask if logs are needed. ### Outcome #### What did you expect? Screen should not glitch while displaying Latex. #### What happened instead? Glitch ### Your phone model Nokia C3 ### Operating system version Android 10 ### Application version and app store 1.3.14 [40103142] (G-b5463) from GPlay Beta ### Homeserver _No response_ ### Will you send logs? No
1.0
Glitch when sending Latex - ### Steps to reproduce 1. Enable Latex in Labs and enable markdown 2. Send any latex formatted code 3. The sceeen stutters and glitches while displaying the latex. Video: https://user-images.githubusercontent.com/62639087/149706093-b0a42bf8-cb15-4c50-a8f4-fd5334eb9ec1.mp4 Please ask if logs are needed. ### Outcome #### What did you expect? Screen should not glitch while displaying Latex. #### What happened instead? Glitch ### Your phone model Nokia C3 ### Operating system version Android 10 ### Application version and app store 1.3.14 [40103142] (G-b5463) from GPlay Beta ### Homeserver _No response_ ### Will you send logs? No
defect
glitch when sending latex steps to reproduce enable latex in labs and enable markdown send any latex formatted code the sceeen stutters and glitches while displaying the latex video please ask if logs are needed outcome what did you expect screen should not glitch while displaying latex what happened instead glitch your phone model nokia operating system version android application version and app store g from gplay beta homeserver no response will you send logs no
1
93,994
10,788,366,287
IssuesEvent
2019-11-05 09:38:57
gpm0009/TFG_MetrominutoWeb
https://api.github.com/repos/gpm0009/TFG_MetrominutoWeb
opened
Cambio nombre milestones
documentation
Los milestones sería mejor que se llamase "Sprint 1", "Sprint 2"... ya que el Sprint conlleva al menos dos reuniones: la inicial y la de retrospectiva al final.
1.0
Cambio nombre milestones - Los milestones sería mejor que se llamase "Sprint 1", "Sprint 2"... ya que el Sprint conlleva al menos dos reuniones: la inicial y la de retrospectiva al final.
non_defect
cambio nombre milestones los milestones sería mejor que se llamase sprint sprint ya que el sprint conlleva al menos dos reuniones la inicial y la de retrospectiva al final
0
58,582
16,612,446,951
IssuesEvent
2021-06-02 13:10:17
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
permalinks only work once
P2 S-Minor T-Defect Z-Chronic Z-Mozilla
If I jump to a specific message in a channel, then I click the down arrow to return to the bottom of the page, then click the link again, Riot doesn't scroll up to the message again
1.0
permalinks only work once - If I jump to a specific message in a channel, then I click the down arrow to return to the bottom of the page, then click the link again, Riot doesn't scroll up to the message again
defect
permalinks only work once if i jump to a specific message in a channel then i click the down arrow to return to the bottom of the page then click the link again riot doesn t scroll up to the message again
1
16,146
2,872,987,402
IssuesEvent
2015-06-08 14:54:15
msimpson/pixelcity
https://api.github.com/repos/msimpson/pixelcity
closed
black screen, no saver
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. copy .scr to C:\windows 2. desktop settings -> screen saver 3. preview What is the expected output? What do you see instead? Instead of the screen saver, I just get a black screen. What version of the product are you using? On what operating system? XP SP2 Please provide any additional information below. The in-screen preview shows a loading screen and then a so-small-you-can't- see-it version of the saver. The full screen version seems to be blank, though. ``` Original issue reported on code.google.com by `wcol...@gmail.com` on 5 May 2009 at 5:36
1.0
black screen, no saver - ``` What steps will reproduce the problem? 1. copy .scr to C:\windows 2. desktop settings -> screen saver 3. preview What is the expected output? What do you see instead? Instead of the screen saver, I just get a black screen. What version of the product are you using? On what operating system? XP SP2 Please provide any additional information below. The in-screen preview shows a loading screen and then a so-small-you-can't- see-it version of the saver. The full screen version seems to be blank, though. ``` Original issue reported on code.google.com by `wcol...@gmail.com` on 5 May 2009 at 5:36
defect
black screen no saver what steps will reproduce the problem copy scr to c windows desktop settings screen saver preview what is the expected output what do you see instead instead of the screen saver i just get a black screen what version of the product are you using on what operating system xp please provide any additional information below the in screen preview shows a loading screen and then a so small you can t see it version of the saver the full screen version seems to be blank though original issue reported on code google com by wcol gmail com on may at
1
332,872
10,112,119,939
IssuesEvent
2019-07-30 14:07:24
our-city-app/mobicage-android-client
https://api.github.com/repos/our-city-app/mobicage-android-client
opened
Addresses should be updatable
priority_critical
Copied from https://github.com/our-city-app/mobicage-ios-client/issues/133 - [ ] Make sure an address + radius can be updated - [ ] Show the address on a map with a pin and a radius
1.0
Addresses should be updatable - Copied from https://github.com/our-city-app/mobicage-ios-client/issues/133 - [ ] Make sure an address + radius can be updated - [ ] Show the address on a map with a pin and a radius
non_defect
addresses should be updatable copied from make sure an address radius can be updated show the address on a map with a pin and a radius
0
52,831
13,225,113,679
IssuesEvent
2020-08-17 20:31:05
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
sim-services/PropagatorServiceUtils::Propagate broken/non-functional (Trac #423)
Migrated from Trac combo reconstruction defect
I'd like to commit the following attached patch with the following commit message: PropagatorServiceUtils::Propagate replaces its input pointer and should thus get it passed by reference. Also the wrong MCTree was modified (the original instead of the output copy). <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/423">https://code.icecube.wisc.edu/projects/icecube/ticket/423</a>, reported by claudio.kopperand owned by olivas</em></summary> <p> ```json { "status": "closed", "changetime": "2012-10-31T17:33:36", "_ts": "1351704816000000", "description": "I'd like to commit the following attached patch with the following commit message:\n\nPropagatorServiceUtils::Propagate replaces its input pointer and should thus get it passed by reference. Also the wrong MCTree was modified (the original instead of the output copy).\n", "reporter": "claudio.kopper", "cc": "", "resolution": "fixed", "time": "2012-06-25T00:59:43", "component": "combo reconstruction", "summary": "sim-services/PropagatorServiceUtils::Propagate broken/non-functional", "priority": "normal", "keywords": "", "milestone": "", "owner": "olivas", "type": "defect" } ``` </p> </details>
1.0
sim-services/PropagatorServiceUtils::Propagate broken/non-functional (Trac #423) - I'd like to commit the following attached patch with the following commit message: PropagatorServiceUtils::Propagate replaces its input pointer and should thus get it passed by reference. Also the wrong MCTree was modified (the original instead of the output copy). <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/423">https://code.icecube.wisc.edu/projects/icecube/ticket/423</a>, reported by claudio.kopperand owned by olivas</em></summary> <p> ```json { "status": "closed", "changetime": "2012-10-31T17:33:36", "_ts": "1351704816000000", "description": "I'd like to commit the following attached patch with the following commit message:\n\nPropagatorServiceUtils::Propagate replaces its input pointer and should thus get it passed by reference. Also the wrong MCTree was modified (the original instead of the output copy).\n", "reporter": "claudio.kopper", "cc": "", "resolution": "fixed", "time": "2012-06-25T00:59:43", "component": "combo reconstruction", "summary": "sim-services/PropagatorServiceUtils::Propagate broken/non-functional", "priority": "normal", "keywords": "", "milestone": "", "owner": "olivas", "type": "defect" } ``` </p> </details>
defect
sim services propagatorserviceutils propagate broken non functional trac i d like to commit the following attached patch with the following commit message propagatorserviceutils propagate replaces its input pointer and should thus get it passed by reference also the wrong mctree was modified the original instead of the output copy migrated from json status closed changetime ts description i d like to commit the following attached patch with the following commit message n npropagatorserviceutils propagate replaces its input pointer and should thus get it passed by reference also the wrong mctree was modified the original instead of the output copy n reporter claudio kopper cc resolution fixed time component combo reconstruction summary sim services propagatorserviceutils propagate broken non functional priority normal keywords milestone owner olivas type defect
1
23,417
7,328,035,426
IssuesEvent
2018-03-04 16:47:22
sans-dfir/sift
https://api.github.com/repos/sans-dfir/sift
closed
ESXI issues
area/builder status/needs-answer
Tried both .ova and also unpacking and adding the .vmdk's and ova to ESXI, getting a "required disk is not present" error followed by "Failed to deploy VM: postNFCData failed: Capacity of uploaded disk is larger than requested" any thoughts?
1.0
ESXI issues - Tried both .ova and also unpacking and adding the .vmdk's and ova to ESXI, getting a "required disk is not present" error followed by "Failed to deploy VM: postNFCData failed: Capacity of uploaded disk is larger than requested" any thoughts?
non_defect
esxi issues tried both ova and also unpacking and adding the vmdk s and ova to esxi getting a required disk is not present error followed by failed to deploy vm postnfcdata failed capacity of uploaded disk is larger than requested any thoughts
0
758,444
26,555,756,628
IssuesEvent
2023-01-20 11:52:30
saudalnasser/starlux
https://api.github.com/repos/saudalnasser/starlux
opened
feat: plugins
type: feature priority: high
## Problem we need a way to allow the users to extend the framework with plugins to the limit where they can almost create their own framework. ## Solution(s) allow users to define a function that will act as the plugin and make the default implementation a plugin.
1.0
feat: plugins - ## Problem we need a way to allow the users to extend the framework with plugins to the limit where they can almost create their own framework. ## Solution(s) allow users to define a function that will act as the plugin and make the default implementation a plugin.
non_defect
feat plugins problem we need a way to allow the users to extend the framework with plugins to the limit where they can almost create their own framework solution s allow users to define a function that will act as the plugin and make the default implementation a plugin
0
17,333
2,999,724,024
IssuesEvent
2015-07-23 20:33:12
dart-lang/sdk
https://api.github.com/repos/dart-lang/sdk
closed
IJ + DAS, no icons in completion dialog
Analyzer-Completion Area-Analyzer Type-Defect
IJ usually displays nice icons next to the completions - this helps with visual search of the completion list. They are missing form the completions model that uses DAS. ![completion](https://cloud.githubusercontent.com/assets/2978182/8633563/26beebf0-27c6-11e5-9ee2-40f9ee4b7103.png)
1.0
IJ + DAS, no icons in completion dialog - IJ usually displays nice icons next to the completions - this helps with visual search of the completion list. They are missing form the completions model that uses DAS. ![completion](https://cloud.githubusercontent.com/assets/2978182/8633563/26beebf0-27c6-11e5-9ee2-40f9ee4b7103.png)
defect
ij das no icons in completion dialog ij usually displays nice icons next to the completions this helps with visual search of the completion list they are missing form the completions model that uses das
1
203,556
23,157,024,798
IssuesEvent
2022-07-29 13:56:48
turkdevops/pusher-js
https://api.github.com/repos/turkdevops/pusher-js
closed
CVE-2022-24771 (High) detected in node-forge-0.9.0.tgz - autoclosed
security vulnerability
## CVE-2022-24771 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.9.0.tgz</b></p></summary> <p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/node-forge/package.json</p> <p> Dependency Hierarchy: - webpack-dev-server-3.11.0.tgz (Root Library) - selfsigned-1.10.7.tgz - :x: **node-forge-0.9.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Forge (also called `node-forge`) is a native implementation of Transport Layer Security in JavaScript. Prior to version 1.3.0, RSA PKCS#1 v1.5 signature verification code is lenient in checking the digest algorithm structure. This can allow a crafted structure that steals padding bytes and uses unchecked portion of the PKCS#1 encoded message to forge a signature when a low public exponent is being used. The issue has been addressed in `node-forge` version 1.3.0. There are currently no known workarounds. <p>Publish Date: 2022-03-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24771>CVE-2022-24771</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24771">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24771</a></p> <p>Release Date: 2022-03-18</p> <p>Fix Resolution (node-forge): 1.3.0</p> <p>Direct dependency fix Resolution (webpack-dev-server): 4.7.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-24771 (High) detected in node-forge-0.9.0.tgz - autoclosed - ## CVE-2022-24771 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.9.0.tgz</b></p></summary> <p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.9.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/node-forge/package.json</p> <p> Dependency Hierarchy: - webpack-dev-server-3.11.0.tgz (Root Library) - selfsigned-1.10.7.tgz - :x: **node-forge-0.9.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Forge (also called `node-forge`) is a native implementation of Transport Layer Security in JavaScript. Prior to version 1.3.0, RSA PKCS#1 v1.5 signature verification code is lenient in checking the digest algorithm structure. This can allow a crafted structure that steals padding bytes and uses unchecked portion of the PKCS#1 encoded message to forge a signature when a low public exponent is being used. The issue has been addressed in `node-forge` version 1.3.0. There are currently no known workarounds. <p>Publish Date: 2022-03-18 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24771>CVE-2022-24771</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24771">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24771</a></p> <p>Release Date: 2022-03-18</p> <p>Fix Resolution (node-forge): 1.3.0</p> <p>Direct dependency fix Resolution (webpack-dev-server): 4.7.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in node forge tgz autoclosed cve high severity vulnerability vulnerable library node forge tgz javascript implementations of network transports cryptography ciphers pki message digests and various utilities library home page a href path to dependency file package json path to vulnerable library node modules node forge package json dependency hierarchy webpack dev server tgz root library selfsigned tgz x node forge tgz vulnerable library found in base branch master vulnerability details forge also called node forge is a native implementation of transport layer security in javascript prior to version rsa pkcs signature verification code is lenient in checking the digest algorithm structure this can allow a crafted structure that steals padding bytes and uses unchecked portion of the pkcs encoded message to forge a signature when a low public exponent is being used the issue has been addressed in node forge version there are currently no known workarounds publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution node forge direct dependency fix resolution webpack dev server step up your open source security game with mend
0
545,139
15,937,089,997
IssuesEvent
2021-04-14 12:01:32
Gird-the-Grid/Grid-the-Grid
https://api.github.com/repos/Gird-the-Grid/Grid-the-Grid
opened
Frontend Grid Management Program
enhancement high priority
* Should be accessible from ControlPanel only to users who completed both company configuration and grid parameters (when `GET`-ing them, `success` is `true`) * Should make call to an API to get electric grid data and display its changes (for now)
1.0
Frontend Grid Management Program - * Should be accessible from ControlPanel only to users who completed both company configuration and grid parameters (when `GET`-ing them, `success` is `true`) * Should make call to an API to get electric grid data and display its changes (for now)
non_defect
frontend grid management program should be accessible from controlpanel only to users who completed both company configuration and grid parameters when get ing them success is true should make call to an api to get electric grid data and display its changes for now
0
17,665
3,012,806,149
IssuesEvent
2015-07-29 02:48:11
yawlfoundation/yawl
https://api.github.com/repos/yawlfoundation/yawl
closed
Editor 20080915 - adding a DTD breaks Decompose to Direct Data Transfer
auto-migrated Component-Editor Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Start a new net. 2. create local-to-net variables. 3. create a task. 4. check for their availability by starting-but-not-doing a Decompose-to-Direct-Data-Transfer, appears OK. 5. add a DTD. 6. create a local-to-net variable of the new type, does NOT appear. 7. create an additional local-to-net standard variable. 8. check for that new var's availability, ALSO does not appear. Can create task variables appropriately, can update parameter mappings manually... but does not appear in Decompose-Direct. Not even if you drop a manually created decomposition and try again. What version of the product are you using? On what operating system? Editor: 2.0b..., 2008-09-15 release. OS: Windows XP SP2 Java: 1.6.0 Please provide any additional information below. ``` Original issue reported on code.google.com by `Trofflup...@gmail.com` on 1 Oct 2008 at 3:14 Attachments: * [Aspec v01 - four basic types, one task.ywl](https://storage.googleapis.com/google-code-attachments/yawl/issue-167/comment-0/Aspec v01 - four basic types, one task.ywl) * [Aspec v02 - add a DTD.ywl](https://storage.googleapis.com/google-code-attachments/yawl/issue-167/comment-0/Aspec v02 - add a DTD.ywl) * [Aspec v03 - add a DTDed variable.ywl](https://storage.googleapis.com/google-code-attachments/yawl/issue-167/comment-0/Aspec v03 - add a DTDed variable.ywl) * [Aspec v04 - added a fifth-basic-sixth-total variable.ywl](https://storage.googleapis.com/google-code-attachments/yawl/issue-167/comment-0/Aspec v04 - added a fifth-basic-sixth-total variable.ywl)
1.0
Editor 20080915 - adding a DTD breaks Decompose to Direct Data Transfer - ``` What steps will reproduce the problem? 1. Start a new net. 2. create local-to-net variables. 3. create a task. 4. check for their availability by starting-but-not-doing a Decompose-to-Direct-Data-Transfer, appears OK. 5. add a DTD. 6. create a local-to-net variable of the new type, does NOT appear. 7. create an additional local-to-net standard variable. 8. check for that new var's availability, ALSO does not appear. Can create task variables appropriately, can update parameter mappings manually... but does not appear in Decompose-Direct. Not even if you drop a manually created decomposition and try again. What version of the product are you using? On what operating system? Editor: 2.0b..., 2008-09-15 release. OS: Windows XP SP2 Java: 1.6.0 Please provide any additional information below. ``` Original issue reported on code.google.com by `Trofflup...@gmail.com` on 1 Oct 2008 at 3:14 Attachments: * [Aspec v01 - four basic types, one task.ywl](https://storage.googleapis.com/google-code-attachments/yawl/issue-167/comment-0/Aspec v01 - four basic types, one task.ywl) * [Aspec v02 - add a DTD.ywl](https://storage.googleapis.com/google-code-attachments/yawl/issue-167/comment-0/Aspec v02 - add a DTD.ywl) * [Aspec v03 - add a DTDed variable.ywl](https://storage.googleapis.com/google-code-attachments/yawl/issue-167/comment-0/Aspec v03 - add a DTDed variable.ywl) * [Aspec v04 - added a fifth-basic-sixth-total variable.ywl](https://storage.googleapis.com/google-code-attachments/yawl/issue-167/comment-0/Aspec v04 - added a fifth-basic-sixth-total variable.ywl)
defect
editor adding a dtd breaks decompose to direct data transfer what steps will reproduce the problem start a new net create local to net variables create a task check for their availability by starting but not doing a decompose to direct data transfer appears ok add a dtd create a local to net variable of the new type does not appear create an additional local to net standard variable check for that new var s availability also does not appear can create task variables appropriately can update parameter mappings manually but does not appear in decompose direct not even if you drop a manually created decomposition and try again what version of the product are you using on what operating system editor release os windows xp java please provide any additional information below original issue reported on code google com by trofflup gmail com on oct at attachments four basic types one task ywl add a dtd ywl add a dtded variable ywl added a fifth basic sixth total variable ywl
1
145,074
13,133,823,777
IssuesEvent
2020-08-06 21:45:34
SergioMorchon/fitbit-views
https://api.github.com/repos/SergioMorchon/fitbit-views
closed
How to disable Phsyical button back(), or pass paremeter with back button press?
documentation question
Hi again, I am having some trouble with parameter passing using the physical back button. I will first explain the situation: I have a view B, which I entered using `next('views/B', {from: 'A');` from view A. This now means that view B expects the parameter "from" whenever entering that view. The problem arises when I leave view B for view C, and when in view C the physical back button is pressed. This causes the regular back(), function to be called which means no parameter is passed back to view B, causing view B to produce the error: App: Unhandled exception: TypeError: wrong type of argument ((null):1,1) App: Unhandled exception: TypeError: Cannot read property 'from' of undefined ? at app/views/B.js Is there someway to do either of the following? 1) Disable the physical back button on some views? 2) Pass parameters in the back() function when the physical back button is pressed? Thanks again for your help!
1.0
How to disable Phsyical button back(), or pass paremeter with back button press? - Hi again, I am having some trouble with parameter passing using the physical back button. I will first explain the situation: I have a view B, which I entered using `next('views/B', {from: 'A');` from view A. This now means that view B expects the parameter "from" whenever entering that view. The problem arises when I leave view B for view C, and when in view C the physical back button is pressed. This causes the regular back(), function to be called which means no parameter is passed back to view B, causing view B to produce the error: App: Unhandled exception: TypeError: wrong type of argument ((null):1,1) App: Unhandled exception: TypeError: Cannot read property 'from' of undefined ? at app/views/B.js Is there someway to do either of the following? 1) Disable the physical back button on some views? 2) Pass parameters in the back() function when the physical back button is pressed? Thanks again for your help!
non_defect
how to disable phsyical button back or pass paremeter with back button press hi again i am having some trouble with parameter passing using the physical back button i will first explain the situation i have a view b which i entered using next views b from a from view a this now means that view b expects the parameter from whenever entering that view the problem arises when i leave view b for view c and when in view c the physical back button is pressed this causes the regular back function to be called which means no parameter is passed back to view b causing view b to produce the error app unhandled exception typeerror wrong type of argument null app unhandled exception typeerror cannot read property from of undefined at app views b js is there someway to do either of the following disable the physical back button on some views pass parameters in the back function when the physical back button is pressed thanks again for your help
0
109,141
13,748,283,812
IssuesEvent
2020-10-06 08:51:35
cyfronet-fid/sat4envi
https://api.github.com/repos/cyfronet-fid/sat4envi
closed
Extended registration form
graphic design
Design new registration form like Copernicus Open Hub with similar fields.
1.0
Extended registration form - Design new registration form like Copernicus Open Hub with similar fields.
non_defect
extended registration form design new registration form like copernicus open hub with similar fields
0
35,426
7,739,207,883
IssuesEvent
2018-05-28 14:51:22
DotJoshJohnson/vscode-xml
https://api.github.com/repos/DotJoshJohnson/vscode-xml
closed
Evaluating xpath - result is undefined
(XPath Evaluator) Defect SemVer: Patch
Very simple xml file open: `<bookstore> <book> <title lang="en">Harry Potter</title> <price>29.99</price> </book> <book> <title lang="en">Learning XML</title> <price>39.95</price> </book> </bookstore>` Evaluate xpath expression, enter: //book/* Result: undefined Same result for every other expression tried.
1.0
Evaluating xpath - result is undefined - Very simple xml file open: `<bookstore> <book> <title lang="en">Harry Potter</title> <price>29.99</price> </book> <book> <title lang="en">Learning XML</title> <price>39.95</price> </book> </bookstore>` Evaluate xpath expression, enter: //book/* Result: undefined Same result for every other expression tried.
defect
evaluating xpath result is undefined very simple xml file open harry potter learning xml evaluate xpath expression enter book result undefined same result for every other expression tried
1
252,145
8,032,477,305
IssuesEvent
2018-07-28 15:52:18
wso2/testgrid
https://api.github.com/repos/wso2/testgrid
opened
Identify the exact infrastructure a given test case has failed
Priority/High Severity/Critical Type/New Feature
**Description:** Using statistical analysis of builds of the test plans in the current iteration, we should try to determine the exact failing infrastructure that lead to test failures/errors. This is much useful than showing lists of failing infrastructure combinations. ![image](https://user-images.githubusercontent.com/936037/43358232-e4cd39e6-92ab-11e8-8ba4-1c9628d28020.png) **Affected Product Version:** m36 **Related Issues:** #906 <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
1.0
Identify the exact infrastructure a given test case has failed - **Description:** Using statistical analysis of builds of the test plans in the current iteration, we should try to determine the exact failing infrastructure that lead to test failures/errors. This is much useful than showing lists of failing infrastructure combinations. ![image](https://user-images.githubusercontent.com/936037/43358232-e4cd39e6-92ab-11e8-8ba4-1c9628d28020.png) **Affected Product Version:** m36 **Related Issues:** #906 <!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. -->
non_defect
identify the exact infrastructure a given test case has failed description using statistical analysis of builds of the test plans in the current iteration we should try to determine the exact failing infrastructure that lead to test failures errors this is much useful than showing lists of failing infrastructure combinations affected product version related issues
0
4,691
2,610,141,092
IssuesEvent
2015-02-26 18:44:28
chrsmith/hedgewars
https://api.github.com/repos/chrsmith/hedgewars
closed
Problem with computer-run team: always does the same action as it does not know what to do
auto-migrated Priority-Low Type-Defect
``` What steps will reproduce the problem? I can't really say: watch the video, the green hedge has already shot once, we're waiting for the second time, but it always does the same actions again (he did it for about thirty seconds, before at least shooting to the right of the screen). What is the expected output? What do you see instead? See video. What version of the product are you using? On what operating system? Hedgewars 0.9.13, Ubuntu 10.10 Please provide any additional information below. ``` ----- Original issue reported on code.google.com by `t.lauxer...@gmail.com` on 29 Nov 2010 at 2:40 * Merged into: #184 Attachments: * [bug_hedgewars.ogv](https://storage.googleapis.com/google-code-attachments/hedgewars/issue-113/comment-0/bug_hedgewars.ogv)
1.0
Problem with computer-run team: always does the same action as it does not know what to do - ``` What steps will reproduce the problem? I can't really say: watch the video, the green hedge has already shot once, we're waiting for the second time, but it always does the same actions again (he did it for about thirty seconds, before at least shooting to the right of the screen). What is the expected output? What do you see instead? See video. What version of the product are you using? On what operating system? Hedgewars 0.9.13, Ubuntu 10.10 Please provide any additional information below. ``` ----- Original issue reported on code.google.com by `t.lauxer...@gmail.com` on 29 Nov 2010 at 2:40 * Merged into: #184 Attachments: * [bug_hedgewars.ogv](https://storage.googleapis.com/google-code-attachments/hedgewars/issue-113/comment-0/bug_hedgewars.ogv)
defect
problem with computer run team always does the same action as it does not know what to do what steps will reproduce the problem i can t really say watch the video the green hedge has already shot once we re waiting for the second time but it always does the same actions again he did it for about thirty seconds before at least shooting to the right of the screen what is the expected output what do you see instead see video what version of the product are you using on what operating system hedgewars ubuntu please provide any additional information below original issue reported on code google com by t lauxer gmail com on nov at merged into attachments
1
85,935
10,697,834,010
IssuesEvent
2019-10-23 17:22:46
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
opened
[Design] Remove gating for Original Claims
526 design vsa-benefits
## User Story or Problem Statement As a _____, I need _____ so I can _____. ## Goal _What outcome(s) do we want to see?_ ## Objectives or Key Results this is meant to further - _lorem ipsum_ ## Acceptance Criteria - [ ] _What needs to happen or be created?_ ## How to configure this issue - [ ] **Attached to a Milestone** (when will this be completed?) - [ ] **Attached to an Epic** (what body of work is this a part of?) - [ ] **Labeled with Team** (`product support`, `analytics-insights`, `operations`, `triage`, `tools-improvements`) - [ ] **Labeled with Practice Area** (`backend`, `frontend`, `devops`, `design`, `research`, `product`, `ia`, `qa`, `analytics`, `call center`, `research`, `accessibility`, `content`) - [ ] **Labeled with Type** (`bug`, `request`, `discovery`, `documentation`, etc.)
1.0
[Design] Remove gating for Original Claims - ## User Story or Problem Statement As a _____, I need _____ so I can _____. ## Goal _What outcome(s) do we want to see?_ ## Objectives or Key Results this is meant to further - _lorem ipsum_ ## Acceptance Criteria - [ ] _What needs to happen or be created?_ ## How to configure this issue - [ ] **Attached to a Milestone** (when will this be completed?) - [ ] **Attached to an Epic** (what body of work is this a part of?) - [ ] **Labeled with Team** (`product support`, `analytics-insights`, `operations`, `triage`, `tools-improvements`) - [ ] **Labeled with Practice Area** (`backend`, `frontend`, `devops`, `design`, `research`, `product`, `ia`, `qa`, `analytics`, `call center`, `research`, `accessibility`, `content`) - [ ] **Labeled with Type** (`bug`, `request`, `discovery`, `documentation`, etc.)
non_defect
remove gating for original claims user story or problem statement as a i need so i can goal what outcome s do we want to see objectives or key results this is meant to further lorem ipsum acceptance criteria what needs to happen or be created how to configure this issue attached to a milestone when will this be completed attached to an epic what body of work is this a part of labeled with team product support analytics insights operations triage tools improvements labeled with practice area backend frontend devops design research product ia qa analytics call center research accessibility content labeled with type bug request discovery documentation etc
0
219,969
16,859,018,102
IssuesEvent
2021-06-21 10:33:09
ahodelin/Bachmann_Archive
https://api.github.com/repos/ahodelin/Bachmann_Archive
opened
Glossary from Google to Wiki
documentation
Hello, I have imported our glossary from Google Doc in our Wiki (https://github.com/ahodelin/Bachmann_Archive/wiki/glossary). **If you want** you can translate something. Thanks
1.0
Glossary from Google to Wiki - Hello, I have imported our glossary from Google Doc in our Wiki (https://github.com/ahodelin/Bachmann_Archive/wiki/glossary). **If you want** you can translate something. Thanks
non_defect
glossary from google to wiki hello i have imported our glossary from google doc in our wiki if you want you can translate something thanks
0
73,130
24,469,915,057
IssuesEvent
2022-10-07 18:43:29
BOINC/boinc
https://api.github.com/repos/BOINC/boinc
closed
VBoxWrapper not quickly updating status info anymore
P: Minor R: worksforme T: Defect C: Apps - VboxWrapper Validate
I crunch long-running tasks for RNA World. Currently, I've got 2 different application versions (which use 2 different vboxwrapper versions), and they behave quite differently! The versions are: - old: vboxwrapper v26167 - new: vboxwrapper v26184 The new application, appears to have some regressions. It is quite a bit more sluggish, as compared to the old. Specifically, the new vboxwrapper: - Only sends an < app_msg_receive > message every 10 seconds, instead of every 1 second like the old app - Only updates the < current_cpu_time > and < fraction_done > within the < app_msg_receive >, about every 90 seconds, instead of every 10 seconds like the old app, thus making the UI not update often. - Takes about half a minute longer to get the < remote_desktop_addr > message, to show the "Show VM Console" button as available. - Takes about 10 seconds longer to exit the processes, when exiting BOINC. Below is a snippet from my Event Log. Slots 17 and 11 are running the old vboxwrapper v26167, while slots 2 and 4 are running the new vboxwrapper v26184. Can we please look into why the new vboxwrapper is so much more sluggish, and maybe fix it? Thanks, Jacob Klein [20160314 VBoxWrapper Sluggishness.txt](https://github.com/BOINC/boinc/files/173163/20160314.VBoxWrapper.Sluggishness.txt)
1.0
VBoxWrapper not quickly updating status info anymore - I crunch long-running tasks for RNA World. Currently, I've got 2 different application versions (which use 2 different vboxwrapper versions), and they behave quite differently! The versions are: - old: vboxwrapper v26167 - new: vboxwrapper v26184 The new application, appears to have some regressions. It is quite a bit more sluggish, as compared to the old. Specifically, the new vboxwrapper: - Only sends an < app_msg_receive > message every 10 seconds, instead of every 1 second like the old app - Only updates the < current_cpu_time > and < fraction_done > within the < app_msg_receive >, about every 90 seconds, instead of every 10 seconds like the old app, thus making the UI not update often. - Takes about half a minute longer to get the < remote_desktop_addr > message, to show the "Show VM Console" button as available. - Takes about 10 seconds longer to exit the processes, when exiting BOINC. Below is a snippet from my Event Log. Slots 17 and 11 are running the old vboxwrapper v26167, while slots 2 and 4 are running the new vboxwrapper v26184. Can we please look into why the new vboxwrapper is so much more sluggish, and maybe fix it? Thanks, Jacob Klein [20160314 VBoxWrapper Sluggishness.txt](https://github.com/BOINC/boinc/files/173163/20160314.VBoxWrapper.Sluggishness.txt)
defect
vboxwrapper not quickly updating status info anymore i crunch long running tasks for rna world currently i ve got different application versions which use different vboxwrapper versions and they behave quite differently the versions are old vboxwrapper new vboxwrapper the new application appears to have some regressions it is quite a bit more sluggish as compared to the old specifically the new vboxwrapper only sends an message every seconds instead of every second like the old app only updates the and within the about every seconds instead of every seconds like the old app thus making the ui not update often takes about half a minute longer to get the message to show the show vm console button as available takes about seconds longer to exit the processes when exiting boinc below is a snippet from my event log slots and are running the old vboxwrapper while slots and are running the new vboxwrapper can we please look into why the new vboxwrapper is so much more sluggish and maybe fix it thanks jacob klein
1
77,152
26,806,449,620
IssuesEvent
2023-02-01 18:41:34
DependencyTrack/dependency-track
https://api.github.com/repos/DependencyTrack/dependency-track
closed
Transitive dependenices are not handled after ingestion of hierarchically merged BOMs
defect
### Current Behavior The Dependency Graph from a hierarchically merged BOM does not show transitive dependencies. ### Steps to Reproduce 1. Upload "bom_foo.xml": ![foo_deps](https://user-images.githubusercontent.com/38717523/214949845-6e83aa56-e7b6-4af8-a6a2-93657960a3c4.png) 2. Upload "bom_bar.xml": ![bar_deps](https://user-images.githubusercontent.com/38717523/214949877-98e93007-8950-4561-b25e-e9d7fe9420ca.png) 4. Hierarchically merge to the `foo` and `bar` BOM: e,g, `docker run --rm -v<bom location>:/tmp/target --name cyclonedx cyclonedx/cyclonedx-cli:0.24.2 merge --input-files /tmp/target/bom_foo.xml /tmp/target/bom_bar.xml --input-format xml --output-file /tmp/target/bom_foobar.xml --output-format xml --hierarchical --group tld.domain --name foobar --version 0.0.1-SNAPSHOT` 5. Upload the resulting "bom_foobar.xml" and observe that only the top-level dependencies are present: ![missing_deps](https://user-images.githubusercontent.com/38717523/214950037-3ddb2459-2ca2-4667-9f48-fd4f585e537f.png) 8. Query the database and to confirm that the `direct_dependencies` information is missing (PGSQL): `SELECT "GROUP", "NAME", "DIRECT_DEPENDENCIES" FROM "COMPONENT" WHERE "PROJECT_ID" IN (SELECT "ID" FROM "PROJECT" WHERE "NAME" LIKE 'bom_foobar')` ![missing_directs](https://user-images.githubusercontent.com/38717523/214950171-95d6e45f-334a-45cf-8dc5-aa03c2289ee7.png) [boms.zip](https://github.com/DependencyTrack/dependency-track/files/10512844/boms.zip) ### Expected Behavior When the result of a hierachical merge is uploaded, one would expect the Dependency Graph to be able to drill down all the dependencies and not just the top level. This _seems_ to be caused by the components no longer all being "top-level" in the BOM after the hierachical merge. Most are now components-of-components. This can be corrected on line 136 of `BomUploadProcessingTask.inform(Event)` by passing `flattenedComponents` instead. So: `ModelConverter.generateDependencies(qm, cycloneDxBom, project, components);` becomes: `ModelConverter.generateDependencies(qm, cycloneDxBom, project, flattenedComponents);` After this change everything seems to work because the transitive `for` loop in `generateDependencies` can now find all the components from the dependencies: ![fixed_deps](https://user-images.githubusercontent.com/38717523/214948498-ec3715f0-f3f2-4378-97ce-39b67f633670.png) Running the previous query also shows the `direct_dependencies` column populated: ![populated_directs](https://user-images.githubusercontent.com/38717523/214948623-99821c4b-cb32-4f9f-b8af-77d4e8de3090.png) Unfortunately I am not knowledgable enough about this code base to fully assess the impact of this change. ### Dependency-Track Version 4.7.0 ### Dependency-Track Distribution Container Image ### Database Server N/A ### Database Server Version _No response_ ### Browser N/A ### Checklist - [X] I have read and understand the [contributing guidelines](https://github.com/DependencyTrack/dependency-track/blob/master/CONTRIBUTING.md#filing-issues) - [X] I have checked the [existing issues](https://github.com/DependencyTrack/dependency-track/issues) for whether this defect was already reported
1.0
Transitive dependenices are not handled after ingestion of hierarchically merged BOMs - ### Current Behavior The Dependency Graph from a hierarchically merged BOM does not show transitive dependencies. ### Steps to Reproduce 1. Upload "bom_foo.xml": ![foo_deps](https://user-images.githubusercontent.com/38717523/214949845-6e83aa56-e7b6-4af8-a6a2-93657960a3c4.png) 2. Upload "bom_bar.xml": ![bar_deps](https://user-images.githubusercontent.com/38717523/214949877-98e93007-8950-4561-b25e-e9d7fe9420ca.png) 4. Hierarchically merge to the `foo` and `bar` BOM: e,g, `docker run --rm -v<bom location>:/tmp/target --name cyclonedx cyclonedx/cyclonedx-cli:0.24.2 merge --input-files /tmp/target/bom_foo.xml /tmp/target/bom_bar.xml --input-format xml --output-file /tmp/target/bom_foobar.xml --output-format xml --hierarchical --group tld.domain --name foobar --version 0.0.1-SNAPSHOT` 5. Upload the resulting "bom_foobar.xml" and observe that only the top-level dependencies are present: ![missing_deps](https://user-images.githubusercontent.com/38717523/214950037-3ddb2459-2ca2-4667-9f48-fd4f585e537f.png) 8. Query the database and to confirm that the `direct_dependencies` information is missing (PGSQL): `SELECT "GROUP", "NAME", "DIRECT_DEPENDENCIES" FROM "COMPONENT" WHERE "PROJECT_ID" IN (SELECT "ID" FROM "PROJECT" WHERE "NAME" LIKE 'bom_foobar')` ![missing_directs](https://user-images.githubusercontent.com/38717523/214950171-95d6e45f-334a-45cf-8dc5-aa03c2289ee7.png) [boms.zip](https://github.com/DependencyTrack/dependency-track/files/10512844/boms.zip) ### Expected Behavior When the result of a hierachical merge is uploaded, one would expect the Dependency Graph to be able to drill down all the dependencies and not just the top level. This _seems_ to be caused by the components no longer all being "top-level" in the BOM after the hierachical merge. Most are now components-of-components. This can be corrected on line 136 of `BomUploadProcessingTask.inform(Event)` by passing `flattenedComponents` instead. So: `ModelConverter.generateDependencies(qm, cycloneDxBom, project, components);` becomes: `ModelConverter.generateDependencies(qm, cycloneDxBom, project, flattenedComponents);` After this change everything seems to work because the transitive `for` loop in `generateDependencies` can now find all the components from the dependencies: ![fixed_deps](https://user-images.githubusercontent.com/38717523/214948498-ec3715f0-f3f2-4378-97ce-39b67f633670.png) Running the previous query also shows the `direct_dependencies` column populated: ![populated_directs](https://user-images.githubusercontent.com/38717523/214948623-99821c4b-cb32-4f9f-b8af-77d4e8de3090.png) Unfortunately I am not knowledgable enough about this code base to fully assess the impact of this change. ### Dependency-Track Version 4.7.0 ### Dependency-Track Distribution Container Image ### Database Server N/A ### Database Server Version _No response_ ### Browser N/A ### Checklist - [X] I have read and understand the [contributing guidelines](https://github.com/DependencyTrack/dependency-track/blob/master/CONTRIBUTING.md#filing-issues) - [X] I have checked the [existing issues](https://github.com/DependencyTrack/dependency-track/issues) for whether this defect was already reported
defect
transitive dependenices are not handled after ingestion of hierarchically merged boms current behavior the dependency graph from a hierarchically merged bom does not show transitive dependencies steps to reproduce upload bom foo xml upload bom bar xml hierarchically merge to the foo and bar bom e g docker run rm v tmp target name cyclonedx cyclonedx cyclonedx cli merge input files tmp target bom foo xml tmp target bom bar xml input format xml output file tmp target bom foobar xml output format xml hierarchical group tld domain name foobar version snapshot upload the resulting bom foobar xml and observe that only the top level dependencies are present query the database and to confirm that the direct dependencies information is missing pgsql select group name direct dependencies from component where project id in select id from project where name like bom foobar expected behavior when the result of a hierachical merge is uploaded one would expect the dependency graph to be able to drill down all the dependencies and not just the top level this seems to be caused by the components no longer all being top level in the bom after the hierachical merge most are now components of components this can be corrected on line of bomuploadprocessingtask inform event by passing flattenedcomponents instead so modelconverter generatedependencies qm cyclonedxbom project components becomes modelconverter generatedependencies qm cyclonedxbom project flattenedcomponents after this change everything seems to work because the transitive for loop in generatedependencies can now find all the components from the dependencies running the previous query also shows the direct dependencies column populated unfortunately i am not knowledgable enough about this code base to fully assess the impact of this change dependency track version dependency track distribution container image database server n a database server version no response browser n a checklist i have read and understand the i have checked the for whether this defect was already reported
1
52,566
13,224,839,166
IssuesEvent
2020-08-17 19:57:19
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
opened
PPC pybindings are broken when ZeroMQ is not present (Trac #2434)
Incomplete Migration Migrated from Trac combo simulation defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2434">https://code.icecube.wisc.edu/projects/icecube/ticket/2434</a>, reported by cweaverand owned by olivas</em></summary> <p> ```json { "status": "accepted", "changetime": "2020-08-06T16:38:34", "_ts": "1596731914086435", "description": "I have been struggling for some time with a python bindings crash when prevents loading any script which transitively loads ppc (e.g. anything using `icecube.simprod.segments`). \n\nThe symptoms look like this:\n{{{\n$ ./env-shell.sh \n************************************************************************\n* *\n* W E L C O M E to I C E T R A Y *\n* *\n* Version combo.trunk r181458 *\n* *\n* You are welcome to visit our Web site *\n* http://icecube.umd.edu *\n* *\n************************************************************************\n\nIcetray environment has:\n I3_SRC = /Users/christopher/Work/IceCube/combo/src\n I3_BUILD = /Users/christopher/Work/IceCube/combo/build\n I3_TESTDATA = /Users/christopher/Work/IceCube/combo/build/test-data\n Python = 2.7.10\nPalladium-II:build $ python\nPython 2.7.10 (default, Feb 7 2017, 00:08:15) \n[GCC 4.2.1 Compatible Apple LLVM 8.0.0 (clang-800.0.34)] on darwin\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\n>>> from icecube import ppc\nTraceback (most recent call last):\n File \"<stdin>\", line 1, in <module>\n File \"/Users/christopher/Work/IceCube/combo/build/lib/icecube/ppc/__init__.py\", line 5, in <module>\n load_pybindings(__name__,__path__)\n File \"/Users/christopher/Work/IceCube/combo/build/lib/icecube/load_pybindings.py\", line 83, in load_pybindings\n m = imp.load_dynamic(name, path[0] + \".so\")\nRuntimeError: extension class wrapper for base class I3CLSimStepToPhotonConverter has not been created yet\n}}}\n\nNote that there's nothing inherently wrong with CLSim itself, and the problem is not fixed by loading CLSim before ppc (which the ppc scripts do correctly, I think):\n{{{\n>>> from icecube import clsim\n>>> dir(clsim)\n['AsyncTap', 'FakeFlasherInfoGenerator', 'FlasherInfoVectToFlasherPulseSeriesConverter', 'GetAntaresOMAcceptance', 'GetAntaresOMAngularSensitivity', 'GetAntaresScatteringCosAngleDistribution', 'GetDefaultParameterizationList', 'GetFlasherParameterizationList', 'GetHybridParameterizationList', 'GetIceCubeDOMAcceptance', 'GetIceCubeDOMAngularSensitivity', 'GetIceCubeFlasherSpectrum', 'GetKM3NeTDOMAcceptance', 'GetPetzoldScatteringCosAngleDistribution', 'I3CLSimEventStatistics', 'I3CLSimFlasherPulse', 'I3CLSimFlasherPulseSeries', 'I3CLSimFunction', 'I3CLSimFunctionAbsLenIceCube', 'I3CLSimFunctionConstant', 'I3CLSimFunctionDeltaPeak', 'I3CLSimFunctionFromTable', 'I3CLSimFunctionMap', 'I3CLSimFunctionPolynomial', 'I3CLSimFunctionRefIndexIceCube', 'I3CLSimFunctionRefIndexQuanFry', 'I3CLSimFunctionScatLenIceCube', 'I3CLSimFunctionScatLenPartic', 'I3CLSimFunctionTester', 'I3CLSimLightSource', 'I3CLSimLightSourceParameterization', 'I3CLSimLightSourceParameterizationSeries', 'I3CLSimLightSourcePropagator', 'I3CLSimLightSourcePropagatorFromI3PropagatorService', 'I3CLSimLightSourcePropagatorSeries', 'I3CLSimLightSourceToStepConverter', 'I3CLSimLightSourceToStepConverterAsync', 'I3CLSimLightSourceToStepConverterFlasher', 'I3CLSimLightSourceToStepConverterPPC', 'I3CLSimMakeHits', 'I3CLSimMakeHitsFromPhotons', 'I3CLSimMakePhotons', 'I3CLSimMediumProperties', 'I3CLSimMediumPropertiesTester', 'I3CLSimOpenCLDevice', 'I3CLSimOpenCLDeviceSeries', 'I3CLSimPhoton', 'I3CLSimPhotonHistory', 'I3CLSimPhotonHistorySeries', 'I3CLSimPhotonSeries', 'I3CLSimPhotonSeriesMap', 'I3CLSimPhotonToMCPEConverter', 'I3CLSimPhotonToMCPEConverterForDOMs', 'I3CLSimRandomDistributionTester', 'I3CLSimRandomNumberGeneratorBenchmark', 'I3CLSimRandomValue', 'I3CLSimRandomValueApplyFunction', 'I3CLSimRandomValueConstant', 'I3CLSimRandomValueFixParameter', 'I3CLSimRandomValueHenyeyGreenstein', 'I3CLSimRandomValueIceCubeFlasherTimeProfile', 'I3CLSimRandomValueInterpolatedDistribution', 'I3CLSimRandomValueMixed', 'I3CLSimRandomValueNormalDistribution', 'I3CLSimRandomValuePtrSeries', 'I3CLSimRandomValueRayleighScatteringCosAngle', 'I3CLSimRandomValueSimplifiedLiu', 'I3CLSimRandomValueUniform', 'I3CLSimRandomValueWlenCherenkovNoDispersion', 'I3CLSimScalarField', 'I3CLSimScalarFieldAnisotropyAbsLenScaling', 'I3CLSimScalarFieldConstant', 'I3CLSimScalarFieldIceTiltZShift', 'I3CLSimScalarFieldTester', 'I3CLSimSimpleGeometry', 'I3CLSimSimpleGeometryFromI3Geometry', 'I3CLSimSimpleGeometryTextFile', 'I3CLSimSpectrumTable', 'I3CLSimStep', 'I3CLSimStepSeries', 'I3CLSimTesterBase', 'I3CLSimVectorTransform', 'I3CLSimVectorTransformConstant', 'I3CLSimVectorTransformMatrix', 'I3CLSimVectorTransformTester', 'MakeAntaresMediumProperties', 'MakeIceCubeMediumProperties', 'MakeIceCubeMediumPropertiesPhotonics', 'NumberOfPhotonsPerMeter', 'PhotonNumberCorrectionFactorAfterBias', 'StandardCandleFlasherPulseSeriesGenerator', '__builtins__', '__doc__', '__file__', '__name__', '__package__', '__path__', 'gammaDistributedNumber', 'initializeOpenCL', 'load_pybindings', 'makeCherenkovWavelengthGenerator', 'makeWavelengthGenerator', 'std_map_indexing_suite_I3CLSimFunctionMap_entry', 'std_map_indexing_suite_I3CLSimPhotonSeriesMap_entry', 'tabulator', 'traysegments', 'util']\n}}}\nIt should be noted, though, that there is no `clsim.I3CLSimStepToPhotonConverter` present. \n\nTHe CLSim cmake output looked like this:\n{{{\n-- + clsim\n-- +-- python [symlinks] \n-- --- ZMQ not found. I3CLSimMakePhotons will not function. \n-- +-- tabulator (have OpenCL 1.2) \n-- +-- safeprimes_base32.gz already downloaded \n-- +-- Geant4 or OpenCL is not installed on your system. clsim will fail if it is not used with parameterizations. \n-- +-- gmp support (make_safeprimes utility) \n-- +-- clsim-pybindings\n}}}\n\nSo, it should not be the case that this is a `BUILD_CLSIM_DATACLASSES_ONLY` build. However, some components are missing due to lack of ZeroMQ. \n\nAdding ZeroMQ to the system then and then rebuilding fixed the problem, but it would be preferable if ppc did not make unfounded assumptions about what symbols will be present in the CLSim libraries. \n", "reporter": "cweaver", "cc": "", "resolution": "", "time": "2020-08-05T02:48:58", "component": "combo simulation", "summary": "PPC pybindings are broken when ZeroMQ is not present", "priority": "normal", "keywords": "", "milestone": "Autumnal Equinox 2020", "owner": "olivas", "type": "defect" } ``` </p> </details>
1.0
PPC pybindings are broken when ZeroMQ is not present (Trac #2434) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2434">https://code.icecube.wisc.edu/projects/icecube/ticket/2434</a>, reported by cweaverand owned by olivas</em></summary> <p> ```json { "status": "accepted", "changetime": "2020-08-06T16:38:34", "_ts": "1596731914086435", "description": "I have been struggling for some time with a python bindings crash when prevents loading any script which transitively loads ppc (e.g. anything using `icecube.simprod.segments`). \n\nThe symptoms look like this:\n{{{\n$ ./env-shell.sh \n************************************************************************\n* *\n* W E L C O M E to I C E T R A Y *\n* *\n* Version combo.trunk r181458 *\n* *\n* You are welcome to visit our Web site *\n* http://icecube.umd.edu *\n* *\n************************************************************************\n\nIcetray environment has:\n I3_SRC = /Users/christopher/Work/IceCube/combo/src\n I3_BUILD = /Users/christopher/Work/IceCube/combo/build\n I3_TESTDATA = /Users/christopher/Work/IceCube/combo/build/test-data\n Python = 2.7.10\nPalladium-II:build $ python\nPython 2.7.10 (default, Feb 7 2017, 00:08:15) \n[GCC 4.2.1 Compatible Apple LLVM 8.0.0 (clang-800.0.34)] on darwin\nType \"help\", \"copyright\", \"credits\" or \"license\" for more information.\n>>> from icecube import ppc\nTraceback (most recent call last):\n File \"<stdin>\", line 1, in <module>\n File \"/Users/christopher/Work/IceCube/combo/build/lib/icecube/ppc/__init__.py\", line 5, in <module>\n load_pybindings(__name__,__path__)\n File \"/Users/christopher/Work/IceCube/combo/build/lib/icecube/load_pybindings.py\", line 83, in load_pybindings\n m = imp.load_dynamic(name, path[0] + \".so\")\nRuntimeError: extension class wrapper for base class I3CLSimStepToPhotonConverter has not been created yet\n}}}\n\nNote that there's nothing inherently wrong with CLSim itself, and the problem is not fixed by loading CLSim before ppc (which the ppc scripts do correctly, I think):\n{{{\n>>> from icecube import clsim\n>>> dir(clsim)\n['AsyncTap', 'FakeFlasherInfoGenerator', 'FlasherInfoVectToFlasherPulseSeriesConverter', 'GetAntaresOMAcceptance', 'GetAntaresOMAngularSensitivity', 'GetAntaresScatteringCosAngleDistribution', 'GetDefaultParameterizationList', 'GetFlasherParameterizationList', 'GetHybridParameterizationList', 'GetIceCubeDOMAcceptance', 'GetIceCubeDOMAngularSensitivity', 'GetIceCubeFlasherSpectrum', 'GetKM3NeTDOMAcceptance', 'GetPetzoldScatteringCosAngleDistribution', 'I3CLSimEventStatistics', 'I3CLSimFlasherPulse', 'I3CLSimFlasherPulseSeries', 'I3CLSimFunction', 'I3CLSimFunctionAbsLenIceCube', 'I3CLSimFunctionConstant', 'I3CLSimFunctionDeltaPeak', 'I3CLSimFunctionFromTable', 'I3CLSimFunctionMap', 'I3CLSimFunctionPolynomial', 'I3CLSimFunctionRefIndexIceCube', 'I3CLSimFunctionRefIndexQuanFry', 'I3CLSimFunctionScatLenIceCube', 'I3CLSimFunctionScatLenPartic', 'I3CLSimFunctionTester', 'I3CLSimLightSource', 'I3CLSimLightSourceParameterization', 'I3CLSimLightSourceParameterizationSeries', 'I3CLSimLightSourcePropagator', 'I3CLSimLightSourcePropagatorFromI3PropagatorService', 'I3CLSimLightSourcePropagatorSeries', 'I3CLSimLightSourceToStepConverter', 'I3CLSimLightSourceToStepConverterAsync', 'I3CLSimLightSourceToStepConverterFlasher', 'I3CLSimLightSourceToStepConverterPPC', 'I3CLSimMakeHits', 'I3CLSimMakeHitsFromPhotons', 'I3CLSimMakePhotons', 'I3CLSimMediumProperties', 'I3CLSimMediumPropertiesTester', 'I3CLSimOpenCLDevice', 'I3CLSimOpenCLDeviceSeries', 'I3CLSimPhoton', 'I3CLSimPhotonHistory', 'I3CLSimPhotonHistorySeries', 'I3CLSimPhotonSeries', 'I3CLSimPhotonSeriesMap', 'I3CLSimPhotonToMCPEConverter', 'I3CLSimPhotonToMCPEConverterForDOMs', 'I3CLSimRandomDistributionTester', 'I3CLSimRandomNumberGeneratorBenchmark', 'I3CLSimRandomValue', 'I3CLSimRandomValueApplyFunction', 'I3CLSimRandomValueConstant', 'I3CLSimRandomValueFixParameter', 'I3CLSimRandomValueHenyeyGreenstein', 'I3CLSimRandomValueIceCubeFlasherTimeProfile', 'I3CLSimRandomValueInterpolatedDistribution', 'I3CLSimRandomValueMixed', 'I3CLSimRandomValueNormalDistribution', 'I3CLSimRandomValuePtrSeries', 'I3CLSimRandomValueRayleighScatteringCosAngle', 'I3CLSimRandomValueSimplifiedLiu', 'I3CLSimRandomValueUniform', 'I3CLSimRandomValueWlenCherenkovNoDispersion', 'I3CLSimScalarField', 'I3CLSimScalarFieldAnisotropyAbsLenScaling', 'I3CLSimScalarFieldConstant', 'I3CLSimScalarFieldIceTiltZShift', 'I3CLSimScalarFieldTester', 'I3CLSimSimpleGeometry', 'I3CLSimSimpleGeometryFromI3Geometry', 'I3CLSimSimpleGeometryTextFile', 'I3CLSimSpectrumTable', 'I3CLSimStep', 'I3CLSimStepSeries', 'I3CLSimTesterBase', 'I3CLSimVectorTransform', 'I3CLSimVectorTransformConstant', 'I3CLSimVectorTransformMatrix', 'I3CLSimVectorTransformTester', 'MakeAntaresMediumProperties', 'MakeIceCubeMediumProperties', 'MakeIceCubeMediumPropertiesPhotonics', 'NumberOfPhotonsPerMeter', 'PhotonNumberCorrectionFactorAfterBias', 'StandardCandleFlasherPulseSeriesGenerator', '__builtins__', '__doc__', '__file__', '__name__', '__package__', '__path__', 'gammaDistributedNumber', 'initializeOpenCL', 'load_pybindings', 'makeCherenkovWavelengthGenerator', 'makeWavelengthGenerator', 'std_map_indexing_suite_I3CLSimFunctionMap_entry', 'std_map_indexing_suite_I3CLSimPhotonSeriesMap_entry', 'tabulator', 'traysegments', 'util']\n}}}\nIt should be noted, though, that there is no `clsim.I3CLSimStepToPhotonConverter` present. \n\nTHe CLSim cmake output looked like this:\n{{{\n-- + clsim\n-- +-- python [symlinks] \n-- --- ZMQ not found. I3CLSimMakePhotons will not function. \n-- +-- tabulator (have OpenCL 1.2) \n-- +-- safeprimes_base32.gz already downloaded \n-- +-- Geant4 or OpenCL is not installed on your system. clsim will fail if it is not used with parameterizations. \n-- +-- gmp support (make_safeprimes utility) \n-- +-- clsim-pybindings\n}}}\n\nSo, it should not be the case that this is a `BUILD_CLSIM_DATACLASSES_ONLY` build. However, some components are missing due to lack of ZeroMQ. \n\nAdding ZeroMQ to the system then and then rebuilding fixed the problem, but it would be preferable if ppc did not make unfounded assumptions about what symbols will be present in the CLSim libraries. \n", "reporter": "cweaver", "cc": "", "resolution": "", "time": "2020-08-05T02:48:58", "component": "combo simulation", "summary": "PPC pybindings are broken when ZeroMQ is not present", "priority": "normal", "keywords": "", "milestone": "Autumnal Equinox 2020", "owner": "olivas", "type": "defect" } ``` </p> </details>
defect
ppc pybindings are broken when zeromq is not present trac migrated from json status accepted changetime ts description i have been struggling for some time with a python bindings crash when prevents loading any script which transitively loads ppc e g anything using icecube simprod segments n nthe symptoms look like this n n env shell sh n n n w e l c o m e to i c e t r a y n n version combo trunk n n you are welcome to visit our web site n n n n nicetray environment has n src users christopher work icecube combo src n build users christopher work icecube combo build n testdata users christopher work icecube combo build test data n python npalladium ii build python npython default feb n on darwin ntype help copyright credits or license for more information n from icecube import ppc ntraceback most recent call last n file line in n file users christopher work icecube combo build lib icecube ppc init py line in n load pybindings name path n file users christopher work icecube combo build lib icecube load pybindings py line in load pybindings n m imp load dynamic name path so nruntimeerror extension class wrapper for base class has not been created yet n n nnote that there s nothing inherently wrong with clsim itself and the problem is not fixed by loading clsim before ppc which the ppc scripts do correctly i think n n from icecube import clsim n dir clsim n n nit should be noted though that there is no clsim present n nthe clsim cmake output looked like this n n clsim n python n zmq not found will not function n tabulator have opencl n safeprimes gz already downloaded n or opencl is not installed on your system clsim will fail if it is not used with parameterizations n gmp support make safeprimes utility n clsim pybindings n n nso it should not be the case that this is a build clsim dataclasses only build however some components are missing due to lack of zeromq n nadding zeromq to the system then and then rebuilding fixed the problem but it would be preferable if ppc did not make unfounded assumptions about what symbols will be present in the clsim libraries n reporter cweaver cc resolution time component combo simulation summary ppc pybindings are broken when zeromq is not present priority normal keywords milestone autumnal equinox owner olivas type defect
1
43,181
11,543,203,417
IssuesEvent
2020-02-18 09:10:19
contao/contao
https://api.github.com/repos/contao/contao
closed
addWizardClass breaks compatibility with older versions
defect
**Affected version(s)** 4.9.0-RC1 **Description** Any extension that uses any picker in a DCA which needed the `wizard` CSS-class will break the styling with 4.9 because of the new `addWizardClass` attribute. I've seen no way to be compatible with 4.4 and 4.9 without distinguishing between the versions within every `eval` of every field. ![image](https://user-images.githubusercontent.com/226890/74236063-a8601880-4cd0-11ea-87d0-9a0efaeb575e.png) > With wizard class from 4.4 ![image](https://user-images.githubusercontent.com/226890/74236194-01c84780-4cd1-11ea-9505-efd9de8fd6ef.png) > Without wizard class as of 4.9 Maybe the handling within https://github.com/contao/contao/blob/4.9/core-bundle/src/Resources/contao/classes/DataContainer.php#L520-L532 could be improved so 4.9 removes the wizard class if we explicitly say `'addWizardClass'=>false`. This way we'd have at least a chance to be backwards compatible.
1.0
addWizardClass breaks compatibility with older versions - **Affected version(s)** 4.9.0-RC1 **Description** Any extension that uses any picker in a DCA which needed the `wizard` CSS-class will break the styling with 4.9 because of the new `addWizardClass` attribute. I've seen no way to be compatible with 4.4 and 4.9 without distinguishing between the versions within every `eval` of every field. ![image](https://user-images.githubusercontent.com/226890/74236063-a8601880-4cd0-11ea-87d0-9a0efaeb575e.png) > With wizard class from 4.4 ![image](https://user-images.githubusercontent.com/226890/74236194-01c84780-4cd1-11ea-9505-efd9de8fd6ef.png) > Without wizard class as of 4.9 Maybe the handling within https://github.com/contao/contao/blob/4.9/core-bundle/src/Resources/contao/classes/DataContainer.php#L520-L532 could be improved so 4.9 removes the wizard class if we explicitly say `'addWizardClass'=>false`. This way we'd have at least a chance to be backwards compatible.
defect
addwizardclass breaks compatibility with older versions affected version s description any extension that uses any picker in a dca which needed the wizard css class will break the styling with because of the new addwizardclass attribute i ve seen no way to be compatible with and without distinguishing between the versions within every eval of every field with wizard class from without wizard class as of maybe the handling within could be improved so removes the wizard class if we explicitly say addwizardclass false this way we d have at least a chance to be backwards compatible
1
756,823
26,487,211,153
IssuesEvent
2023-01-17 19:06:03
cds-snc/notification-planning
https://api.github.com/repos/cds-snc/notification-planning
closed
Design: Make limits more visible
High Priority | Haute priorité UX Dev
_Make limits more visible _ ## Description As a Notify Sender, I need to be able to track how many text messages I've sent today so that I can plan to send more sms messages within my limits. As a Notify Sender, I need to be able to track how many text messages I have left for the day so that I don't get cut off from sending text message notifications mid-job. Respecting our infrastructure capacity. Show daily and yearly limits on the dashboard page, with warning indicator when >=XX% [ADMIN] Prototype A (expose fragments) ??? - [ ] design prototype for testing with group A of users - [ ] expose sms fragments in from notifications table to dashboard - [ ] when creating a template, "this template is going to be # fragments long" - [ ] expose before they send, at the review stage how many fragments you will be sending so users can plan better by having this information earlier in the flow - [ ] more accurate communication of limits (more limits than Prototype B) and better prediction of limits for us - [ ] The SMS limit applies but operational on notification fragments. Prototype B (no fragments) - [ ] design prototype for testing with group B of users - [ ] Add character counting (with some limitations) to template editing stage - [ ] Use SMS character limit of 612 including variables - [ ] Show tip for the counter: template length will increase once/if variables are added - [ ] Reject template if it's over 612 characters when saving - [ ] If template ends up being larger than 612 characters when ready to send, it will fail and user will get an error message if variables are too long - [ ] Dashboard show how many text messages have been sent, how many are left for the day - [ ] The SMS limit applies but operational on notifications (not fragments). ## Acceptance Criteria** (Definition of done) Given some context, when (X) action occurs, then (Y) outcome is achieved - [ ] Error message should match character counter * A11y * Bilingualism * Privacy considerations * Security controls in place * Measuring success and metrics ## QA Steps - [ ] Tested in a realistic production scenario
1.0
Design: Make limits more visible - _Make limits more visible _ ## Description As a Notify Sender, I need to be able to track how many text messages I've sent today so that I can plan to send more sms messages within my limits. As a Notify Sender, I need to be able to track how many text messages I have left for the day so that I don't get cut off from sending text message notifications mid-job. Respecting our infrastructure capacity. Show daily and yearly limits on the dashboard page, with warning indicator when >=XX% [ADMIN] Prototype A (expose fragments) ??? - [ ] design prototype for testing with group A of users - [ ] expose sms fragments in from notifications table to dashboard - [ ] when creating a template, "this template is going to be # fragments long" - [ ] expose before they send, at the review stage how many fragments you will be sending so users can plan better by having this information earlier in the flow - [ ] more accurate communication of limits (more limits than Prototype B) and better prediction of limits for us - [ ] The SMS limit applies but operational on notification fragments. Prototype B (no fragments) - [ ] design prototype for testing with group B of users - [ ] Add character counting (with some limitations) to template editing stage - [ ] Use SMS character limit of 612 including variables - [ ] Show tip for the counter: template length will increase once/if variables are added - [ ] Reject template if it's over 612 characters when saving - [ ] If template ends up being larger than 612 characters when ready to send, it will fail and user will get an error message if variables are too long - [ ] Dashboard show how many text messages have been sent, how many are left for the day - [ ] The SMS limit applies but operational on notifications (not fragments). ## Acceptance Criteria** (Definition of done) Given some context, when (X) action occurs, then (Y) outcome is achieved - [ ] Error message should match character counter * A11y * Bilingualism * Privacy considerations * Security controls in place * Measuring success and metrics ## QA Steps - [ ] Tested in a realistic production scenario
non_defect
design make limits more visible make limits more visible description as a notify sender i need to be able to track how many text messages i ve sent today so that i can plan to send more sms messages within my limits as a notify sender i need to be able to track how many text messages i have left for the day so that i don t get cut off from sending text message notifications mid job respecting our infrastructure capacity show daily and yearly limits on the dashboard page with warning indicator when xx prototype a expose fragments design prototype for testing with group a of users expose sms fragments in from notifications table to dashboard when creating a template this template is going to be fragments long expose before they send at the review stage how many fragments you will be sending so users can plan better by having this information earlier in the flow more accurate communication of limits more limits than prototype b and better prediction of limits for us the sms limit applies but operational on notification fragments prototype b no fragments design prototype for testing with group b of users add character counting with some limitations to template editing stage use sms character limit of including variables show tip for the counter template length will increase once if variables are added reject template if it s over characters when saving if template ends up being larger than characters when ready to send it will fail and user will get an error message if variables are too long dashboard show how many text messages have been sent how many are left for the day the sms limit applies but operational on notifications not fragments acceptance criteria definition of done given some context when x action occurs then y outcome is achieved error message should match character counter bilingualism privacy considerations security controls in place measuring success and metrics qa steps tested in a realistic production scenario
0
24,321
3,963,426,648
IssuesEvent
2016-05-02 20:24:24
opencaching/opencaching-pl
https://api.github.com/repos/opencaching/opencaching-pl
closed
Automatyczne tłumaczenia opisów nie działają
Component_Cache Priority_Low Type_Defect
Jeśli przełączymy język strony na inny niż polski, to zamiast opisu strony jest tylko: > Automatic translation thanks to: > Technologia Google™ I nic poza tym. Jeśli Google pokręciło i nie da się z tym nic zrobić, to już lepiej zostawić oryginalny opis.
1.0
Automatyczne tłumaczenia opisów nie działają - Jeśli przełączymy język strony na inny niż polski, to zamiast opisu strony jest tylko: > Automatic translation thanks to: > Technologia Google™ I nic poza tym. Jeśli Google pokręciło i nie da się z tym nic zrobić, to już lepiej zostawić oryginalny opis.
defect
automatyczne tłumaczenia opisów nie działają jeśli przełączymy język strony na inny niż polski to zamiast opisu strony jest tylko automatic translation thanks to technologia google™ i nic poza tym jeśli google pokręciło i nie da się z tym nic zrobić to już lepiej zostawić oryginalny opis
1
105,138
9,030,036,858
IssuesEvent
2019-02-08 01:38:10
sourcegraph/sourcegraph
https://api.github.com/repos/sourcegraph/sourcegraph
closed
Test plan for automatic Postgres upgrades in the server image
test-plan
We'd like to test both the documentation and the automated upgrades. Go to https://github.com/sourcegraph/sourcegraph/blob/master/doc/admin/migration/3_0.md#bundled-postgres-upgrade and test automated and manual upgrades to `3.0.1-rc.3` coming from `2.13.6`, `3.0.0` and ~`3.0.0-beta5`~ `3.0.0-beta.4`. You can create a fresh install of each of these versions by running the correspondent Docker container. Ensure you clean-up your data directory between each test (e.g. `rm -rf ~/.sourcegraph`) Here's a check-list for you to copy into a comment in this issue and tick as you go: ``` - [ ] Test manual upgrade from `2.13.6` to `3.0.1-rc.3`. - [ ] Test manual upgrade from `3.0.0` to `3.0.1-rc.3`. - [ ] Test automatic upgrade from `2.13.6` to `3.0.1-rc.3`. - [ ] Test automatic upgrade from `3.0.0-beta.4` to `3.0.1-rc.3`. - [ ] Test automatic upgrade from `3.0.0` to `3.0.1-rc.3`. ``` Part of #1404
1.0
Test plan for automatic Postgres upgrades in the server image - We'd like to test both the documentation and the automated upgrades. Go to https://github.com/sourcegraph/sourcegraph/blob/master/doc/admin/migration/3_0.md#bundled-postgres-upgrade and test automated and manual upgrades to `3.0.1-rc.3` coming from `2.13.6`, `3.0.0` and ~`3.0.0-beta5`~ `3.0.0-beta.4`. You can create a fresh install of each of these versions by running the correspondent Docker container. Ensure you clean-up your data directory between each test (e.g. `rm -rf ~/.sourcegraph`) Here's a check-list for you to copy into a comment in this issue and tick as you go: ``` - [ ] Test manual upgrade from `2.13.6` to `3.0.1-rc.3`. - [ ] Test manual upgrade from `3.0.0` to `3.0.1-rc.3`. - [ ] Test automatic upgrade from `2.13.6` to `3.0.1-rc.3`. - [ ] Test automatic upgrade from `3.0.0-beta.4` to `3.0.1-rc.3`. - [ ] Test automatic upgrade from `3.0.0` to `3.0.1-rc.3`. ``` Part of #1404
non_defect
test plan for automatic postgres upgrades in the server image we d like to test both the documentation and the automated upgrades go to and test automated and manual upgrades to rc coming from and beta you can create a fresh install of each of these versions by running the correspondent docker container ensure you clean up your data directory between each test e g rm rf sourcegraph here s a check list for you to copy into a comment in this issue and tick as you go test manual upgrade from to rc test manual upgrade from to rc test automatic upgrade from to rc test automatic upgrade from beta to rc test automatic upgrade from to rc part of
0
31,925
4,306,123,489
IssuesEvent
2016-07-21 01:00:59
medic/medic-webapp
https://api.github.com/repos/medic/medic-webapp
opened
Percentage target values are confusing
1 - Scheduled Needs Design Work UI/UX
Percentage type targets show progress towards the target as a percentage which makes it hard to understand. ![target-percentage-old](https://cloud.githubusercontent.com/assets/2064938/17008162/f0b4280e-4f42-11e6-86f1-658103842181.png) In this example it's not clear whether you're 50% or 45% of the way to the target. One possible improvement is to state the progress as "30 of 60" instead of "50%".
1.0
Percentage target values are confusing - Percentage type targets show progress towards the target as a percentage which makes it hard to understand. ![target-percentage-old](https://cloud.githubusercontent.com/assets/2064938/17008162/f0b4280e-4f42-11e6-86f1-658103842181.png) In this example it's not clear whether you're 50% or 45% of the way to the target. One possible improvement is to state the progress as "30 of 60" instead of "50%".
non_defect
percentage target values are confusing percentage type targets show progress towards the target as a percentage which makes it hard to understand in this example it s not clear whether you re or of the way to the target one possible improvement is to state the progress as of instead of
0
16,520
10,523,579,632
IssuesEvent
2019-09-30 11:21:53
kyma-project/helm-broker
https://api.github.com/repos/kyma-project/helm-broker
closed
Improve controller readiness probe
area/service-catalog enhancement
<!-- Thank you for your contribution. Before you submit the issue: 1. Search open and closed issues for duplicates. 2. Read the contributing guidelines. --> **Description** The `helm-controller` readiness probe is implemented in the dummy way to just return 200OK on every call. We should return 200OK only when the controller is able to work. We need to find the best way to implement the checking of the controller status. **Ideas:** - In SBU controller we do the test reconcile with the empty object and we expect it to process in order to indicate the readiness of the controller. - We can wait until the `controller-runtime` probes implementation is ready and use it. https://github.com/kubernetes-sigs/controller-runtime/pull/419
1.0
Improve controller readiness probe - <!-- Thank you for your contribution. Before you submit the issue: 1. Search open and closed issues for duplicates. 2. Read the contributing guidelines. --> **Description** The `helm-controller` readiness probe is implemented in the dummy way to just return 200OK on every call. We should return 200OK only when the controller is able to work. We need to find the best way to implement the checking of the controller status. **Ideas:** - In SBU controller we do the test reconcile with the empty object and we expect it to process in order to indicate the readiness of the controller. - We can wait until the `controller-runtime` probes implementation is ready and use it. https://github.com/kubernetes-sigs/controller-runtime/pull/419
non_defect
improve controller readiness probe thank you for your contribution before you submit the issue search open and closed issues for duplicates read the contributing guidelines description the helm controller readiness probe is implemented in the dummy way to just return on every call we should return only when the controller is able to work we need to find the best way to implement the checking of the controller status ideas in sbu controller we do the test reconcile with the empty object and we expect it to process in order to indicate the readiness of the controller we can wait until the controller runtime probes implementation is ready and use it
0
60,592
17,023,465,914
IssuesEvent
2021-07-03 02:10:26
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
church gray or brown?
Component: mapnik Priority: minor Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 2.47pm, Wednesday, 19th August 2009]** area tagged with amenity = place_of_worship religion = christian rendered as a gray area with a cross. But: building = yes amenity = place_of_worship religion = christian rendered as a brown (red) area with a cross
1.0
church gray or brown? - **[Submitted to the original trac issue database at 2.47pm, Wednesday, 19th August 2009]** area tagged with amenity = place_of_worship religion = christian rendered as a gray area with a cross. But: building = yes amenity = place_of_worship religion = christian rendered as a brown (red) area with a cross
defect
church gray or brown area tagged with amenity place of worship religion christian rendered as a gray area with a cross but building yes amenity place of worship religion christian rendered as a brown red area with a cross
1
179,717
21,580,304,877
IssuesEvent
2022-05-02 17:58:47
vincenzodistasio97/excel-to-json
https://api.github.com/repos/vincenzodistasio97/excel-to-json
opened
CVE-2020-7598 (Medium) detected in minimist-1.2.0.tgz, minimist-0.0.8.tgz
security vulnerability
## CVE-2020-7598 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>minimist-1.2.0.tgz</b>, <b>minimist-0.0.8.tgz</b></p></summary> <p> <details><summary><b>minimist-1.2.0.tgz</b></p></summary> <p>parse argument options</p> <p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz">https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz</a></p> <p>Path to dependency file: /client/package.json</p> <p>Path to vulnerable library: /client/node_modules/meow/node_modules/minimist/package.json,/client/node_modules/rc/node_modules/minimist/package.json,/client/node_modules/sane/node_modules/minimist/package.json,/client/node_modules/cosmiconfig/node_modules/minimist/package.json</p> <p> Dependency Hierarchy: - react-scripts-1.1.1.tgz (Root Library) - postcss-loader-2.0.8.tgz - postcss-load-config-1.2.0.tgz - cosmiconfig-2.2.2.tgz - :x: **minimist-1.2.0.tgz** (Vulnerable Library) </details> <details><summary><b>minimist-0.0.8.tgz</b></p></summary> <p>parse argument options</p> <p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz">https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz</a></p> <p> Dependency Hierarchy: - react-scripts-1.1.1.tgz (Root Library) - babel-loader-7.1.2.tgz - mkdirp-0.5.1.tgz - :x: **minimist-0.0.8.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/vincenzodistasio97/excel-to-json/commit/e367d4db4134dc676344b2b9fb2443300bd3c9c7">e367d4db4134dc676344b2b9fb2443300bd3c9c7</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> minimist before 1.2.2 could be tricked into adding or modifying properties of Object.prototype using a "constructor" or "__proto__" payload. <p>Publish Date: 2020-03-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7598>CVE-2020-7598</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/substack/minimist/commit/63e7ed05aa4b1889ec2f3b196426db4500cbda94">https://github.com/substack/minimist/commit/63e7ed05aa4b1889ec2f3b196426db4500cbda94</a></p> <p>Release Date: 2020-03-11</p> <p>Fix Resolution (minimist): 1.2.3</p> <p>Direct dependency fix Resolution (react-scripts): 1.1.2</p><p>Fix Resolution (minimist): 0.2.1</p> <p>Direct dependency fix Resolution (react-scripts): 1.1.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-7598 (Medium) detected in minimist-1.2.0.tgz, minimist-0.0.8.tgz - ## CVE-2020-7598 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>minimist-1.2.0.tgz</b>, <b>minimist-0.0.8.tgz</b></p></summary> <p> <details><summary><b>minimist-1.2.0.tgz</b></p></summary> <p>parse argument options</p> <p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz">https://registry.npmjs.org/minimist/-/minimist-1.2.0.tgz</a></p> <p>Path to dependency file: /client/package.json</p> <p>Path to vulnerable library: /client/node_modules/meow/node_modules/minimist/package.json,/client/node_modules/rc/node_modules/minimist/package.json,/client/node_modules/sane/node_modules/minimist/package.json,/client/node_modules/cosmiconfig/node_modules/minimist/package.json</p> <p> Dependency Hierarchy: - react-scripts-1.1.1.tgz (Root Library) - postcss-loader-2.0.8.tgz - postcss-load-config-1.2.0.tgz - cosmiconfig-2.2.2.tgz - :x: **minimist-1.2.0.tgz** (Vulnerable Library) </details> <details><summary><b>minimist-0.0.8.tgz</b></p></summary> <p>parse argument options</p> <p>Library home page: <a href="https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz">https://registry.npmjs.org/minimist/-/minimist-0.0.8.tgz</a></p> <p> Dependency Hierarchy: - react-scripts-1.1.1.tgz (Root Library) - babel-loader-7.1.2.tgz - mkdirp-0.5.1.tgz - :x: **minimist-0.0.8.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/vincenzodistasio97/excel-to-json/commit/e367d4db4134dc676344b2b9fb2443300bd3c9c7">e367d4db4134dc676344b2b9fb2443300bd3c9c7</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> minimist before 1.2.2 could be tricked into adding or modifying properties of Object.prototype using a "constructor" or "__proto__" payload. <p>Publish Date: 2020-03-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7598>CVE-2020-7598</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/substack/minimist/commit/63e7ed05aa4b1889ec2f3b196426db4500cbda94">https://github.com/substack/minimist/commit/63e7ed05aa4b1889ec2f3b196426db4500cbda94</a></p> <p>Release Date: 2020-03-11</p> <p>Fix Resolution (minimist): 1.2.3</p> <p>Direct dependency fix Resolution (react-scripts): 1.1.2</p><p>Fix Resolution (minimist): 0.2.1</p> <p>Direct dependency fix Resolution (react-scripts): 1.1.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in minimist tgz minimist tgz cve medium severity vulnerability vulnerable libraries minimist tgz minimist tgz minimist tgz parse argument options library home page a href path to dependency file client package json path to vulnerable library client node modules meow node modules minimist package json client node modules rc node modules minimist package json client node modules sane node modules minimist package json client node modules cosmiconfig node modules minimist package json dependency hierarchy react scripts tgz root library postcss loader tgz postcss load config tgz cosmiconfig tgz x minimist tgz vulnerable library minimist tgz parse argument options library home page a href dependency hierarchy react scripts tgz root library babel loader tgz mkdirp tgz x minimist tgz vulnerable library found in head commit a href found in base branch master vulnerability details minimist before could be tricked into adding or modifying properties of object prototype using a constructor or proto payload publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution minimist direct dependency fix resolution react scripts fix resolution minimist direct dependency fix resolution react scripts step up your open source security game with whitesource
0
82,925
7,856,607,840
IssuesEvent
2018-06-21 08:10:25
hyrise/hyrise
https://api.github.com/repos/hyrise/hyrise
closed
Benchmarks fail with Cuckoo hash error
Bug Test/Benchmark
The benchmarks (running on Athen for the Speedcenter) fail somewhat randomly (but a lot) with the following error, during the JoinHash Benchmarks: `src/lib/utils/cuckoo_hashtable.hpp:93 There is a cycle in Cuckoo. Need to rehash with different hash functions` If that's helpful, here the parameters for the benchmarks we're running: ``` /var/www/codespeed/repo/build-release/hyriseOperatorBenchmark --benchmark_format=console --benchmark_out=/var/www/codespeed/repo/last_benchmark.json --benchmark_out_format=json --benchmark_repetitions=5 --benchmark_report_aggregates_only=true ```
1.0
Benchmarks fail with Cuckoo hash error - The benchmarks (running on Athen for the Speedcenter) fail somewhat randomly (but a lot) with the following error, during the JoinHash Benchmarks: `src/lib/utils/cuckoo_hashtable.hpp:93 There is a cycle in Cuckoo. Need to rehash with different hash functions` If that's helpful, here the parameters for the benchmarks we're running: ``` /var/www/codespeed/repo/build-release/hyriseOperatorBenchmark --benchmark_format=console --benchmark_out=/var/www/codespeed/repo/last_benchmark.json --benchmark_out_format=json --benchmark_repetitions=5 --benchmark_report_aggregates_only=true ```
non_defect
benchmarks fail with cuckoo hash error the benchmarks running on athen for the speedcenter fail somewhat randomly but a lot with the following error during the joinhash benchmarks src lib utils cuckoo hashtable hpp there is a cycle in cuckoo need to rehash with different hash functions if that s helpful here the parameters for the benchmarks we re running var www codespeed repo build release hyriseoperatorbenchmark benchmark format console benchmark out var www codespeed repo last benchmark json benchmark out format json benchmark repetitions benchmark report aggregates only true
0
31,078
6,423,853,855
IssuesEvent
2017-08-09 12:10:40
wooowooo/phpsocks5
https://api.github.com/repos/wooowooo/phpsocks5
closed
出現錯誤 Create table 1 error.
auto-migrated Priority-Medium Type-Defect
``` socks5.php 開頭 $dbhost = 'free-mysql.BizHostNet.com'; $dbport = '3306'; $dbuser = '1328439894'; $dbpass = '****'; $dbname = '1328439894'; 進入 socks5.php 提示錯誤 Create table 1 error. 服務器的 LOG FriAMCSTE_RFebruaryC1123 process 1 FriAMCSTE_RFebruaryC1123 process 2 75fa68619c0f3967547a735202d572d5 FriAMCSTE_RFebruaryC1123 process 3 75fa68619c0f3967547a735202d572d5 FriAMCSTE_RFebruaryC1123 process 4 75fa68619c0f3967547a735202d572d5 FriAMCSTE_RFebruaryC1123 process 5 75fa68619c0f3967547a735202d572d5 before decrypt postdata: () FriAMCSTE_RFebruaryC1123 process 6 75fa68619c0f3967547a735202d572d5 after decrypt postdata: () FriAMCSTE_RFebruaryC1123 create table process 主機 Apache/2.0.55 (Unix) PHP/4.4.5 Server 資料庫使用 MySQL 大致上訊息為上面所示,希望能幫忙查看哪邊有問題 ``` Original issue reported on code.google.com by `sony.and...@gmail.com` on 16 Feb 2012 at 8:33
1.0
出現錯誤 Create table 1 error. - ``` socks5.php 開頭 $dbhost = 'free-mysql.BizHostNet.com'; $dbport = '3306'; $dbuser = '1328439894'; $dbpass = '****'; $dbname = '1328439894'; 進入 socks5.php 提示錯誤 Create table 1 error. 服務器的 LOG FriAMCSTE_RFebruaryC1123 process 1 FriAMCSTE_RFebruaryC1123 process 2 75fa68619c0f3967547a735202d572d5 FriAMCSTE_RFebruaryC1123 process 3 75fa68619c0f3967547a735202d572d5 FriAMCSTE_RFebruaryC1123 process 4 75fa68619c0f3967547a735202d572d5 FriAMCSTE_RFebruaryC1123 process 5 75fa68619c0f3967547a735202d572d5 before decrypt postdata: () FriAMCSTE_RFebruaryC1123 process 6 75fa68619c0f3967547a735202d572d5 after decrypt postdata: () FriAMCSTE_RFebruaryC1123 create table process 主機 Apache/2.0.55 (Unix) PHP/4.4.5 Server 資料庫使用 MySQL 大致上訊息為上面所示,希望能幫忙查看哪邊有問題 ``` Original issue reported on code.google.com by `sony.and...@gmail.com` on 16 Feb 2012 at 8:33
defect
出現錯誤 create table error php 開頭 dbhost free mysql bizhostnet com dbport dbuser dbpass dbname 進入 php 提示錯誤 create table error 服務器的 log friamcste process friamcste process friamcste process friamcste process friamcste process before decrypt postdata friamcste process after decrypt postdata friamcste create table process 主機 apache unix php server 資料庫使用 mysql 大致上訊息為上面所示,希望能幫忙查看哪邊有問題 original issue reported on code google com by sony and gmail com on feb at
1
114,800
17,262,657,299
IssuesEvent
2021-07-22 09:45:42
lukebroganws/NuGetGallery
https://api.github.com/repos/lukebroganws/NuGetGallery
opened
WS-2017-3772 (High) detected in underscore.string-3.3.5.tgz
security vulnerability
## WS-2017-3772 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>underscore.string-3.3.5.tgz</b></p></summary> <p>String manipulation extensions for Underscore.js javascript library.</p> <p>Library home page: <a href="https://registry.npmjs.org/underscore.string/-/underscore.string-3.3.5.tgz">https://registry.npmjs.org/underscore.string/-/underscore.string-3.3.5.tgz</a></p> <p>Path to dependency file: NuGetGallery/src/Bootstrap/package.json</p> <p>Path to vulnerable library: NuGetGallery/src/Bootstrap/node_modules/underscore.string/package.json</p> <p> Dependency Hierarchy: - grunt-1.4.1.tgz (Root Library) - grunt-legacy-util-2.0.1.tgz - :x: **underscore.string-3.3.5.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebroganws/NuGetGallery/commit/297b0beab28e2208dead800ba93269fc067c6932">297b0beab28e2208dead800ba93269fc067c6932</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Regular Expression Denial of Service (ReDoS) vulnerability was found in underscore.string 2.4.0 through 3.3.5. <p>Publish Date: 2017-09-08 <p>URL: <a href=https://github.com/esamattis/underscore.string/issues/510>WS-2017-3772</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"underscore.string","packageVersion":"3.3.5","packageFilePaths":["/src/Bootstrap/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt:1.4.1;grunt-legacy-util:2.0.1;underscore.string:3.3.5","isMinimumFixVersionAvailable":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"WS-2017-3772","vulnerabilityDetails":"Regular Expression Denial of Service (ReDoS) vulnerability was found in underscore.string 2.4.0 through 3.3.5.","vulnerabilityUrl":"https://github.com/esamattis/underscore.string/issues/510","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
WS-2017-3772 (High) detected in underscore.string-3.3.5.tgz - ## WS-2017-3772 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>underscore.string-3.3.5.tgz</b></p></summary> <p>String manipulation extensions for Underscore.js javascript library.</p> <p>Library home page: <a href="https://registry.npmjs.org/underscore.string/-/underscore.string-3.3.5.tgz">https://registry.npmjs.org/underscore.string/-/underscore.string-3.3.5.tgz</a></p> <p>Path to dependency file: NuGetGallery/src/Bootstrap/package.json</p> <p>Path to vulnerable library: NuGetGallery/src/Bootstrap/node_modules/underscore.string/package.json</p> <p> Dependency Hierarchy: - grunt-1.4.1.tgz (Root Library) - grunt-legacy-util-2.0.1.tgz - :x: **underscore.string-3.3.5.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/lukebroganws/NuGetGallery/commit/297b0beab28e2208dead800ba93269fc067c6932">297b0beab28e2208dead800ba93269fc067c6932</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Regular Expression Denial of Service (ReDoS) vulnerability was found in underscore.string 2.4.0 through 3.3.5. <p>Publish Date: 2017-09-08 <p>URL: <a href=https://github.com/esamattis/underscore.string/issues/510>WS-2017-3772</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"underscore.string","packageVersion":"3.3.5","packageFilePaths":["/src/Bootstrap/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt:1.4.1;grunt-legacy-util:2.0.1;underscore.string:3.3.5","isMinimumFixVersionAvailable":false}],"baseBranches":["main"],"vulnerabilityIdentifier":"WS-2017-3772","vulnerabilityDetails":"Regular Expression Denial of Service (ReDoS) vulnerability was found in underscore.string 2.4.0 through 3.3.5.","vulnerabilityUrl":"https://github.com/esamattis/underscore.string/issues/510","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_defect
ws high detected in underscore string tgz ws high severity vulnerability vulnerable library underscore string tgz string manipulation extensions for underscore js javascript library library home page a href path to dependency file nugetgallery src bootstrap package json path to vulnerable library nugetgallery src bootstrap node modules underscore string package json dependency hierarchy grunt tgz root library grunt legacy util tgz x underscore string tgz vulnerable library found in head commit a href found in base branch main vulnerability details regular expression denial of service redos vulnerability was found in underscore string through publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree grunt grunt legacy util underscore string isminimumfixversionavailable false basebranches vulnerabilityidentifier ws vulnerabilitydetails regular expression denial of service redos vulnerability was found in underscore string through vulnerabilityurl
0
58,533
16,591,127,229
IssuesEvent
2021-06-01 07:52:09
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
closed
Manual section using-jooq-with-jdbctemplate shouldn't use deprecated JdbcTemplate method
C: Documentation E: All Editions P: Medium T: Defect
The jdbcTemplate query method in this example is deprecated in Spring v5.x in favor of its varargs equivelant. . See https://github.com/spring-projects/spring-framework/wiki/Upgrading-to-Spring-Framework-5.x#data-access-and-transactions It can be replaced with the new one(switch second and third arguments). It won't make a big difference other than not seeing the deprecated strike-through fonts after coping the example code into your IDE which is a little bit confusing.
1.0
Manual section using-jooq-with-jdbctemplate shouldn't use deprecated JdbcTemplate method - The jdbcTemplate query method in this example is deprecated in Spring v5.x in favor of its varargs equivelant. . See https://github.com/spring-projects/spring-framework/wiki/Upgrading-to-Spring-Framework-5.x#data-access-and-transactions It can be replaced with the new one(switch second and third arguments). It won't make a big difference other than not seeing the deprecated strike-through fonts after coping the example code into your IDE which is a little bit confusing.
defect
manual section using jooq with jdbctemplate shouldn t use deprecated jdbctemplate method the jdbctemplate query method in this example is deprecated in spring x in favor of its varargs equivelant see it can be replaced with the new one switch second and third arguments it won t make a big difference other than not seeing the deprecated strike through fonts after coping the example code into your ide which is a little bit confusing
1
31,240
6,472,668,794
IssuesEvent
2017-08-17 14:26:18
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
[discovery] multicast discovery doesn't work after loading some data
Team: Core Type: Defect
I have two nodes with default configs. If I start one node, then after some time start another - everything works. But if I start one node, load it with some data (5000 entries to one map, overall 865 kbytes) and then start another node - they can't see each other. Nothing is shown in logs, no attempts to connect, second node starts as if the first node wasn't running at all. I tried increasing `multicast-timeout-seconds` to 10 seconds - no effect. Update: tried switching to TCP/IP discovery - bug is not reproduced
1.0
[discovery] multicast discovery doesn't work after loading some data - I have two nodes with default configs. If I start one node, then after some time start another - everything works. But if I start one node, load it with some data (5000 entries to one map, overall 865 kbytes) and then start another node - they can't see each other. Nothing is shown in logs, no attempts to connect, second node starts as if the first node wasn't running at all. I tried increasing `multicast-timeout-seconds` to 10 seconds - no effect. Update: tried switching to TCP/IP discovery - bug is not reproduced
defect
multicast discovery doesn t work after loading some data i have two nodes with default configs if i start one node then after some time start another everything works but if i start one node load it with some data entries to one map overall kbytes and then start another node they can t see each other nothing is shown in logs no attempts to connect second node starts as if the first node wasn t running at all i tried increasing multicast timeout seconds to seconds no effect update tried switching to tcp ip discovery bug is not reproduced
1
55,635
14,599,438,675
IssuesEvent
2020-12-21 04:12:05
SAP/fundamental-ngx
https://api.github.com/repos/SAP/fundamental-ngx
closed
Multi Input: Selected one list from the dropdown but in the token it showing two items added
Defect Hunting High bug platform
Description: Selected one list from the dropdown but in the token it showing two items added Steps: - Type 2 letter it showing the auto suggestion - Select one item from the list. - In the token it is added 2 and also it showing al Expected: In th token it should be one item Screenshot attached. ![image](https://user-images.githubusercontent.com/32538291/100314643-187c2880-2fdd-11eb-9eea-511006bd4d30.png) Expected: ![image](https://user-images.githubusercontent.com/32538291/100314721-482b3080-2fdd-11eb-9885-5e1a47facdd9.png)
1.0
Multi Input: Selected one list from the dropdown but in the token it showing two items added - Description: Selected one list from the dropdown but in the token it showing two items added Steps: - Type 2 letter it showing the auto suggestion - Select one item from the list. - In the token it is added 2 and also it showing al Expected: In th token it should be one item Screenshot attached. ![image](https://user-images.githubusercontent.com/32538291/100314643-187c2880-2fdd-11eb-9eea-511006bd4d30.png) Expected: ![image](https://user-images.githubusercontent.com/32538291/100314721-482b3080-2fdd-11eb-9885-5e1a47facdd9.png)
defect
multi input selected one list from the dropdown but in the token it showing two items added description selected one list from the dropdown but in the token it showing two items added steps type letter it showing the auto suggestion select one item from the list in the token it is added and also it showing al expected in th token it should be one item screenshot attached expected
1
23,501
4,020,603,530
IssuesEvent
2016-05-16 18:59:32
AdamsLair/duality
https://api.github.com/repos/AdamsLair/duality
opened
Reduce Struct Array Serialization Verbosity
Core Task Unit-Test This
#### Summary In the course of issue #279, it became apparent that both XML and Binary serialization write a lot of unnecessary information when serializing a large amount of structs. Investigate ways to reduce the amount of written data in the case of struct and struct array serialization. #### Analysis - While `null` values can be serialized easily, there currently is no special case for `default(T)` struct values. They could be omitted safely just as well. - `XmlSerializer`: Just don't populate the current struct element in `WriteStruct`. - `BinarySerializer`: Use the skip-field array to skip all fields. - Note: This is only viable for non-custom serialization. - Theoretically, this could be extended towards omitting `default(T)` fields of a struct value. - Doesn't work for classes, as they can have arbitrary constructors, so their field's default is unknown. - May result in less readable XML. Could be considered for binary serialization. - Add unit tests that check specifically for potential problems with the omission of values. #### Attachments Sample excerpt of verbose XML struct array serialization: ```xml <data dataType="Array" type="Duality.Plugins.Tilemaps.TileInput[]" id="295733828" length="556"> <item dataType="Struct" type="Duality.Plugins.Tilemaps.TileInput"> <Collision dataType="Struct" type="Duality.Plugins.Tilemaps.TileCollisionShapes"> <Layer0 dataType="Enum" type="Duality.Plugins.Tilemaps.TileCollisionShape" name="Free" value="0" /> <Layer1 dataType="Enum" type="Duality.Plugins.Tilemaps.TileCollisionShape" name="Free" value="0" /> <Layer2 dataType="Enum" type="Duality.Plugins.Tilemaps.TileCollisionShape" name="Free" value="0" /> <Layer3 dataType="Enum" type="Duality.Plugins.Tilemaps.TileCollisionShape" name="Free" value="0" /> </Collision> <DepthOffset dataType="Int">0</DepthOffset> <IsVertical dataType="Bool">false</IsVertical> </item> <item dataType="Struct" type="Duality.Plugins.Tilemaps.TileInput"> <Collision dataType="Struct" type="Duality.Plugins.Tilemaps.TileCollisionShapes"> <Layer0 dataType="Enum" type="Duality.Plugins.Tilemaps.TileCollisionShape" name="Free" value="0" /> <Layer1 dataType="Enum" type="Duality.Plugins.Tilemaps.TileCollisionShape" name="Free" value="0" /> <Layer2 dataType="Enum" type="Duality.Plugins.Tilemaps.TileCollisionShape" name="Free" value="0" /> <Layer3 dataType="Enum" type="Duality.Plugins.Tilemaps.TileCollisionShape" name="Free" value="0" /> </Collision> <DepthOffset dataType="Int">0</DepthOffset> <IsVertical dataType="Bool">false</IsVertical> </item> <!-- etc. --> </data> ```
1.0
Reduce Struct Array Serialization Verbosity - #### Summary In the course of issue #279, it became apparent that both XML and Binary serialization write a lot of unnecessary information when serializing a large amount of structs. Investigate ways to reduce the amount of written data in the case of struct and struct array serialization. #### Analysis - While `null` values can be serialized easily, there currently is no special case for `default(T)` struct values. They could be omitted safely just as well. - `XmlSerializer`: Just don't populate the current struct element in `WriteStruct`. - `BinarySerializer`: Use the skip-field array to skip all fields. - Note: This is only viable for non-custom serialization. - Theoretically, this could be extended towards omitting `default(T)` fields of a struct value. - Doesn't work for classes, as they can have arbitrary constructors, so their field's default is unknown. - May result in less readable XML. Could be considered for binary serialization. - Add unit tests that check specifically for potential problems with the omission of values. #### Attachments Sample excerpt of verbose XML struct array serialization: ```xml <data dataType="Array" type="Duality.Plugins.Tilemaps.TileInput[]" id="295733828" length="556"> <item dataType="Struct" type="Duality.Plugins.Tilemaps.TileInput"> <Collision dataType="Struct" type="Duality.Plugins.Tilemaps.TileCollisionShapes"> <Layer0 dataType="Enum" type="Duality.Plugins.Tilemaps.TileCollisionShape" name="Free" value="0" /> <Layer1 dataType="Enum" type="Duality.Plugins.Tilemaps.TileCollisionShape" name="Free" value="0" /> <Layer2 dataType="Enum" type="Duality.Plugins.Tilemaps.TileCollisionShape" name="Free" value="0" /> <Layer3 dataType="Enum" type="Duality.Plugins.Tilemaps.TileCollisionShape" name="Free" value="0" /> </Collision> <DepthOffset dataType="Int">0</DepthOffset> <IsVertical dataType="Bool">false</IsVertical> </item> <item dataType="Struct" type="Duality.Plugins.Tilemaps.TileInput"> <Collision dataType="Struct" type="Duality.Plugins.Tilemaps.TileCollisionShapes"> <Layer0 dataType="Enum" type="Duality.Plugins.Tilemaps.TileCollisionShape" name="Free" value="0" /> <Layer1 dataType="Enum" type="Duality.Plugins.Tilemaps.TileCollisionShape" name="Free" value="0" /> <Layer2 dataType="Enum" type="Duality.Plugins.Tilemaps.TileCollisionShape" name="Free" value="0" /> <Layer3 dataType="Enum" type="Duality.Plugins.Tilemaps.TileCollisionShape" name="Free" value="0" /> </Collision> <DepthOffset dataType="Int">0</DepthOffset> <IsVertical dataType="Bool">false</IsVertical> </item> <!-- etc. --> </data> ```
non_defect
reduce struct array serialization verbosity summary in the course of issue it became apparent that both xml and binary serialization write a lot of unnecessary information when serializing a large amount of structs investigate ways to reduce the amount of written data in the case of struct and struct array serialization analysis while null values can be serialized easily there currently is no special case for default t struct values they could be omitted safely just as well xmlserializer just don t populate the current struct element in writestruct binaryserializer use the skip field array to skip all fields note this is only viable for non custom serialization theoretically this could be extended towards omitting default t fields of a struct value doesn t work for classes as they can have arbitrary constructors so their field s default is unknown may result in less readable xml could be considered for binary serialization add unit tests that check specifically for potential problems with the omission of values attachments sample excerpt of verbose xml struct array serialization xml false false
0
4,988
2,610,163,496
IssuesEvent
2015-02-26 18:51:51
chrsmith/republic-at-war
https://api.github.com/repos/chrsmith/republic-at-war
closed
Text
auto-migrated Priority-Medium Type-Defect
``` GC bonus with the icon of a vengence frigate engine (speed boost of some kind?) displays [MISSING] for both title and description ``` ----- Original issue reported on code.google.com by `z3r0...@gmail.com` on 3 May 2011 at 6:42
1.0
Text - ``` GC bonus with the icon of a vengence frigate engine (speed boost of some kind?) displays [MISSING] for both title and description ``` ----- Original issue reported on code.google.com by `z3r0...@gmail.com` on 3 May 2011 at 6:42
defect
text gc bonus with the icon of a vengence frigate engine speed boost of some kind displays for both title and description original issue reported on code google com by gmail com on may at
1
142,422
13,024,024,225
IssuesEvent
2020-07-27 11:04:23
Xceptance/neodymium-library
https://api.github.com/repos/Xceptance/neodymium-library
closed
Research GitHub Actions by setting up a plugin project
Low Priority documentation improvement task
Create file: `.github/workflows/maven.yml` Fill with configuration: ```YML name: Java CI on: push: branches: - develop jobs: build: runs-on: ubuntu-latest steps: - uses: actions/checkout@v1 - name: Set up JDK 1.8 uses: actions/setup-java@v1 with: java-version: 1.8 - name: Build with Maven run: mvn -B clean compile package -P release-neodymium ```
1.0
Research GitHub Actions by setting up a plugin project - Create file: `.github/workflows/maven.yml` Fill with configuration: ```YML name: Java CI on: push: branches: - develop jobs: build: runs-on: ubuntu-latest steps: - uses: actions/checkout@v1 - name: Set up JDK 1.8 uses: actions/setup-java@v1 with: java-version: 1.8 - name: Build with Maven run: mvn -B clean compile package -P release-neodymium ```
non_defect
research github actions by setting up a plugin project create file github workflows maven yml fill with configuration yml name java ci on push branches develop jobs build runs on ubuntu latest steps uses actions checkout name set up jdk uses actions setup java with java version name build with maven run mvn b clean compile package p release neodymium
0
73,738
24,782,382,093
IssuesEvent
2022-10-24 06:50:24
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Cannot persistently clear unread messages in rooms containing threads
T-Defect S-Major O-Uncommon A-Threads Z-Labs Z-ThreadsNotifications
### Steps to reproduce Load my account in our test server. See this <img width="380" alt="Screenshot 2022-02-17 at 13 19 10" src="https://user-images.githubusercontent.com/57377/154489855-a049b003-3811-4e57-bb71-fda2ac9a5243.png"> Click on all of the rooms in the room list to clear unread <img width="315" alt="Screenshot 2022-02-17 at 13 19 29" src="https://user-images.githubusercontent.com/57377/154489945-31b206f1-fa46-47a4-b090-bb1537e803e9.png"> Reload the page and the unreads are back <img width="310" alt="Screenshot 2022-02-17 at 13 20 02" src="https://user-images.githubusercontent.com/57377/154489982-d7163242-f5b4-46d0-95fd-567fcea98ab5.png"> ### Outcome #### What did you expect? For the unread markers and badges to not return after they were cleared #### What happened instead? They came back ### Operating system OS X ### Browser information Chrome ### URL for webapp develop.element.io ### Application version Element version: ba55473a0e4e-react-c19aa957b655-js-74d24f38f742 Olm version: 3.2.8 ### Homeserver threads-dev.lab.element.dev ### Will you send logs? No
1.0
Cannot persistently clear unread messages in rooms containing threads - ### Steps to reproduce Load my account in our test server. See this <img width="380" alt="Screenshot 2022-02-17 at 13 19 10" src="https://user-images.githubusercontent.com/57377/154489855-a049b003-3811-4e57-bb71-fda2ac9a5243.png"> Click on all of the rooms in the room list to clear unread <img width="315" alt="Screenshot 2022-02-17 at 13 19 29" src="https://user-images.githubusercontent.com/57377/154489945-31b206f1-fa46-47a4-b090-bb1537e803e9.png"> Reload the page and the unreads are back <img width="310" alt="Screenshot 2022-02-17 at 13 20 02" src="https://user-images.githubusercontent.com/57377/154489982-d7163242-f5b4-46d0-95fd-567fcea98ab5.png"> ### Outcome #### What did you expect? For the unread markers and badges to not return after they were cleared #### What happened instead? They came back ### Operating system OS X ### Browser information Chrome ### URL for webapp develop.element.io ### Application version Element version: ba55473a0e4e-react-c19aa957b655-js-74d24f38f742 Olm version: 3.2.8 ### Homeserver threads-dev.lab.element.dev ### Will you send logs? No
defect
cannot persistently clear unread messages in rooms containing threads steps to reproduce load my account in our test server see this img width alt screenshot at src click on all of the rooms in the room list to clear unread img width alt screenshot at src reload the page and the unreads are back img width alt screenshot at src outcome what did you expect for the unread markers and badges to not return after they were cleared what happened instead they came back operating system os x browser information chrome url for webapp develop element io application version element version react js olm version homeserver threads dev lab element dev will you send logs no
1
9,035
12,130,107,896
IssuesEvent
2020-04-23 00:30:39
GoogleCloudPlatform/python-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
closed
remove gcp-devrel-py-tools from appengine/standard/analytics/requirements-test.txt
priority: p2 remove-gcp-devrel-py-tools type: process
remove gcp-devrel-py-tools from appengine/standard/analytics/requirements-test.txt
1.0
remove gcp-devrel-py-tools from appengine/standard/analytics/requirements-test.txt - remove gcp-devrel-py-tools from appengine/standard/analytics/requirements-test.txt
non_defect
remove gcp devrel py tools from appengine standard analytics requirements test txt remove gcp devrel py tools from appengine standard analytics requirements test txt
0
56,378
15,047,074,200
IssuesEvent
2021-02-03 08:24:47
primefaces/primereact
https://api.github.com/repos/primefaces/primereact
closed
SplitterPanel is not exported in TypeScript
defect
**I'm submitting a ...** (check one with "x") ``` [ x] bug report [ ] feature request [ ] support request => Please do not submit support request here, instead see https://forum.primefaces.org/viewforum.php?f=57 ``` **Current behavior** When trying to use Splitter with typescript, the export for SplitterPanel is missing from the type definitions import { Splitter, SplitterPanel } from "primereact/splitter"; **Expected behavior** SplitterPanel can be imported from the "primereact/splitter" module * **React version:** ^17.0.1 * **PrimeReact version:** ^6.0.1 * **Language:** [TypeScript 6.14.10]
1.0
SplitterPanel is not exported in TypeScript - **I'm submitting a ...** (check one with "x") ``` [ x] bug report [ ] feature request [ ] support request => Please do not submit support request here, instead see https://forum.primefaces.org/viewforum.php?f=57 ``` **Current behavior** When trying to use Splitter with typescript, the export for SplitterPanel is missing from the type definitions import { Splitter, SplitterPanel } from "primereact/splitter"; **Expected behavior** SplitterPanel can be imported from the "primereact/splitter" module * **React version:** ^17.0.1 * **PrimeReact version:** ^6.0.1 * **Language:** [TypeScript 6.14.10]
defect
splitterpanel is not exported in typescript i m submitting a check one with x bug report feature request support request please do not submit support request here instead see current behavior when trying to use splitter with typescript the export for splitterpanel is missing from the type definitions import splitter splitterpanel from primereact splitter expected behavior splitterpanel can be imported from the primereact splitter module react version primereact version language
1
42,503
9,222,164,135
IssuesEvent
2019-03-11 21:57:20
joomla/joomla-cms
https://api.github.com/repos/joomla/joomla-cms
closed
[4.0] SCSS Build script strange behavior
No Code Attached Yet
### Steps to reproduce the issue When tinkering around with #24110 I got one strange behavior: I moved the ```media_source/plg_editors_tinymce/css/tinymce-builder.css``` to ```media_source/plg_editors_tinymce/scss/tinymce-builder.scss``` and want to benefit from the automatic build tool. To my surprise the file is never converted to ```media/plg_editors_tinymce/css/tinymce-builder.css``` So I looked into [compilecss.es6.js](https://github.com/joomla/joomla-cms/blob/4.0-dev/build/build-modules-js/compilecss.es6.js) and (at least for me) are some logical flaws in this file: #### Overview I understand this file, that it loops through the ```media_source``` folder (+ subfolder) and collects all files which are [not in the blacklist](https://github.com/joomla/joomla-cms/blob/4.0-dev/build/build-modules-js/compilecss.es6.js#L61) (I think we should add here ```*.gif``` and ```*.html```, too). ##### CSS [If the file is a ```CSS``` file](https://github.com/joomla/joomla-cms/blob/4.0-dev/build/build-modules-js/compilecss.es6.js#L68), it will be copied to the media folder, minified etc. => works well ##### SCSS ###### **Problem 1** [If the file is a ```SCSS``` file](https://github.com/joomla/joomla-cms/blob/4.0-dev/build/build-modules-js/compilecss.es6.js#L65) and not starts with an underscore (```_```) the file will be added to a collection ```files```. Here is the first bug: the test should be: ```if (file.match(/\.scss$/) && !file.match(/(\/|\\)_[^\/\\]+$/)) {``` because in ```file``` is the whole path and not the filename only. So we should check for ```/``` or ```\``` followed by an ```_``` and then no ```/``` or ```\``` until the end to get wrong filenames. ###### **Problem 2** The next problem here is, that the [SCSS compiling](https://github.com/joomla/joomla-cms/blob/4.0-dev/build/build-modules-js/compilecss.es6.js#L91-L93) is in the [folder loop](https://github.com/joomla/joomla-cms/blob/4.0-dev/build/build-modules-js/compilecss.es6.js#L60), so the added ```files``` will be executed again and again (if we have more folders), it should be moved one line down, outside of the loop. ###### **Problem 3** No let's assume, the SCSS compiling works (it does not reliable, see Problem 4), then we have the following scenario: The SCSS and CSS files will be collected, looped and SCSS compiles + CSS copied. But SCSS files has to be copied, too, when they are converted to CSS files (at least the new CSS files has to). But the moment, [SCSS are compiled to the new CSS files](https://github.com/joomla/joomla-cms/blob/4.0-dev/build/build-modules-js/compilecss.es6.js#L92), [they never appear in the copy area](https://github.com/joomla/joomla-cms/blob/4.0-dev/build/build-modules-js/compilecss.es6.js#L68-L77), because they're not in the collection. So the solution would be, to first collect all SCSS files, compile them and then collect the CSS files + copy them. ###### **Problem 4** In theory the SCSS files get collected and then compiled to CSS, but when I executed ```npm i```my SCSS files (and others like [client.scss](https://github.com/joomla/joomla-cms/blob/4.0-dev/build/media_source/plg_installer_webinstaller/scss/client.scss)) was not compiled. So I added some console-debug-outputs and was very surprised, that [the ```CompileScss.compile(...)``` function](https://github.com/joomla/joomla-cms/blob/4.0-dev/build/build-modules-js/compilecss.es6.js#L92) was called before [the collection of the SCSS files](https://github.com/joomla/joomla-cms/blob/4.0-dev/build/build-modules-js/compilecss.es6.js#L66). So only [the predefined files](https://github.com/joomla/joomla-cms/blob/4.0-dev/build/build-modules-js/compilecss.es6.js#L42-L51) are compiled. My assumption would be, that [the crawling of the folder structure](https://github.com/joomla/joomla-cms/blob/4.0-dev/build/build-modules-js/compilecss.es6.js#L61) takes much longer than the other stuff and if I remember correctly, different calls are not executed in a sequence (if not enforeced by something like ["sequence"](https://www.npmjs.com/package/sequence)). So it calls the ```Recurse```function and continues in the script, while in the background the folders are crawled. So the ```compileSCSS``` will executed before the crawling is finished. #### GIST [Here is a first GIST](https://gist.github.com/bembelimen/817abf9078ba40d6f4c6460d2d517740) which solves 1+2+3 but still fails on problem 4, but perhaps it helps? @dgrammatiko
1.0
[4.0] SCSS Build script strange behavior - ### Steps to reproduce the issue When tinkering around with #24110 I got one strange behavior: I moved the ```media_source/plg_editors_tinymce/css/tinymce-builder.css``` to ```media_source/plg_editors_tinymce/scss/tinymce-builder.scss``` and want to benefit from the automatic build tool. To my surprise the file is never converted to ```media/plg_editors_tinymce/css/tinymce-builder.css``` So I looked into [compilecss.es6.js](https://github.com/joomla/joomla-cms/blob/4.0-dev/build/build-modules-js/compilecss.es6.js) and (at least for me) are some logical flaws in this file: #### Overview I understand this file, that it loops through the ```media_source``` folder (+ subfolder) and collects all files which are [not in the blacklist](https://github.com/joomla/joomla-cms/blob/4.0-dev/build/build-modules-js/compilecss.es6.js#L61) (I think we should add here ```*.gif``` and ```*.html```, too). ##### CSS [If the file is a ```CSS``` file](https://github.com/joomla/joomla-cms/blob/4.0-dev/build/build-modules-js/compilecss.es6.js#L68), it will be copied to the media folder, minified etc. => works well ##### SCSS ###### **Problem 1** [If the file is a ```SCSS``` file](https://github.com/joomla/joomla-cms/blob/4.0-dev/build/build-modules-js/compilecss.es6.js#L65) and not starts with an underscore (```_```) the file will be added to a collection ```files```. Here is the first bug: the test should be: ```if (file.match(/\.scss$/) && !file.match(/(\/|\\)_[^\/\\]+$/)) {``` because in ```file``` is the whole path and not the filename only. So we should check for ```/``` or ```\``` followed by an ```_``` and then no ```/``` or ```\``` until the end to get wrong filenames. ###### **Problem 2** The next problem here is, that the [SCSS compiling](https://github.com/joomla/joomla-cms/blob/4.0-dev/build/build-modules-js/compilecss.es6.js#L91-L93) is in the [folder loop](https://github.com/joomla/joomla-cms/blob/4.0-dev/build/build-modules-js/compilecss.es6.js#L60), so the added ```files``` will be executed again and again (if we have more folders), it should be moved one line down, outside of the loop. ###### **Problem 3** No let's assume, the SCSS compiling works (it does not reliable, see Problem 4), then we have the following scenario: The SCSS and CSS files will be collected, looped and SCSS compiles + CSS copied. But SCSS files has to be copied, too, when they are converted to CSS files (at least the new CSS files has to). But the moment, [SCSS are compiled to the new CSS files](https://github.com/joomla/joomla-cms/blob/4.0-dev/build/build-modules-js/compilecss.es6.js#L92), [they never appear in the copy area](https://github.com/joomla/joomla-cms/blob/4.0-dev/build/build-modules-js/compilecss.es6.js#L68-L77), because they're not in the collection. So the solution would be, to first collect all SCSS files, compile them and then collect the CSS files + copy them. ###### **Problem 4** In theory the SCSS files get collected and then compiled to CSS, but when I executed ```npm i```my SCSS files (and others like [client.scss](https://github.com/joomla/joomla-cms/blob/4.0-dev/build/media_source/plg_installer_webinstaller/scss/client.scss)) was not compiled. So I added some console-debug-outputs and was very surprised, that [the ```CompileScss.compile(...)``` function](https://github.com/joomla/joomla-cms/blob/4.0-dev/build/build-modules-js/compilecss.es6.js#L92) was called before [the collection of the SCSS files](https://github.com/joomla/joomla-cms/blob/4.0-dev/build/build-modules-js/compilecss.es6.js#L66). So only [the predefined files](https://github.com/joomla/joomla-cms/blob/4.0-dev/build/build-modules-js/compilecss.es6.js#L42-L51) are compiled. My assumption would be, that [the crawling of the folder structure](https://github.com/joomla/joomla-cms/blob/4.0-dev/build/build-modules-js/compilecss.es6.js#L61) takes much longer than the other stuff and if I remember correctly, different calls are not executed in a sequence (if not enforeced by something like ["sequence"](https://www.npmjs.com/package/sequence)). So it calls the ```Recurse```function and continues in the script, while in the background the folders are crawled. So the ```compileSCSS``` will executed before the crawling is finished. #### GIST [Here is a first GIST](https://gist.github.com/bembelimen/817abf9078ba40d6f4c6460d2d517740) which solves 1+2+3 but still fails on problem 4, but perhaps it helps? @dgrammatiko
non_defect
scss build script strange behavior steps to reproduce the issue when tinkering around with i got one strange behavior i moved the media source plg editors tinymce css tinymce builder css to media source plg editors tinymce scss tinymce builder scss and want to benefit from the automatic build tool to my surprise the file is never converted to media plg editors tinymce css tinymce builder css so i looked into and at least for me are some logical flaws in this file overview i understand this file that it loops through the media source folder subfolder and collects all files which are i think we should add here gif and html too css it will be copied to the media folder minified etc works well scss problem and not starts with an underscore the file will be added to a collection files here is the first bug the test should be if file match scss file match because in file is the whole path and not the filename only so we should check for or followed by an and then no or until the end to get wrong filenames problem the next problem here is that the is in the so the added files will be executed again and again if we have more folders it should be moved one line down outside of the loop problem no let s assume the scss compiling works it does not reliable see problem then we have the following scenario the scss and css files will be collected looped and scss compiles css copied but scss files has to be copied too when they are converted to css files at least the new css files has to but the moment because they re not in the collection so the solution would be to first collect all scss files compile them and then collect the css files copy them problem in theory the scss files get collected and then compiled to css but when i executed npm i my scss files and others like was not compiled so i added some console debug outputs and was very surprised that was called before so only are compiled my assumption would be that takes much longer than the other stuff and if i remember correctly different calls are not executed in a sequence if not enforeced by something like so it calls the recurse function and continues in the script while in the background the folders are crawled so the compilescss will executed before the crawling is finished gist which solves but still fails on problem but perhaps it helps dgrammatiko
0
61,095
12,145,941,944
IssuesEvent
2020-04-24 10:12:30
strangerstudios/pmpro-register-helper
https://api.github.com/repos/strangerstudios/pmpro-register-helper
closed
Some field types don't respect the class attribute
Difficulty: Easy Impact: Low Status: Needs Code
radio buttons, grouped check boxes, and hidden fields should add a class attribute to the main html element in the getHTML method. https://github.com/strangerstudios/pmpro-register-helper/blob/dev/classes/class.field.php#L403 A work around is to use the divclass property which adds the class to the wrapping div.
1.0
Some field types don't respect the class attribute - radio buttons, grouped check boxes, and hidden fields should add a class attribute to the main html element in the getHTML method. https://github.com/strangerstudios/pmpro-register-helper/blob/dev/classes/class.field.php#L403 A work around is to use the divclass property which adds the class to the wrapping div.
non_defect
some field types don t respect the class attribute radio buttons grouped check boxes and hidden fields should add a class attribute to the main html element in the gethtml method a work around is to use the divclass property which adds the class to the wrapping div
0
36,336
7,891,900,836
IssuesEvent
2018-06-28 13:36:23
emory-libraries/Primo-releases
https://api.github.com/repos/emory-libraries/Primo-releases
opened
INC02935709 Legacy ETD Embargo restrictions not current in DiscoverE (see examples in description)
defect
SCO staff noticed that revised embargo restrictions in Legacy ETDs are not showing up in DiscoverE. Two examples: 1.) Title in DiscoverE: Protecting Infants through Tdap Vaccination during Pregnancy: A Qualitative Analysis of the Perspectives of Obstetrician-Gynecologists; Legacy ETD link: https://legacy-etd.library.emory.edu/view/record/pid/emory:rm7xf; 2) Title in DiscoverE: Imagined Places: politics and narratives in a disputed indo-tibetan borderland; Legacy ETD link: https://legacy-etd.library.emory.edu/view/record/pid/emory:f3jbp. I am wondering if this has anything to do with the recent url change to Legacy ETDs that occurred in August 2017 when we deployed New ETDs?
1.0
INC02935709 Legacy ETD Embargo restrictions not current in DiscoverE (see examples in description) - SCO staff noticed that revised embargo restrictions in Legacy ETDs are not showing up in DiscoverE. Two examples: 1.) Title in DiscoverE: Protecting Infants through Tdap Vaccination during Pregnancy: A Qualitative Analysis of the Perspectives of Obstetrician-Gynecologists; Legacy ETD link: https://legacy-etd.library.emory.edu/view/record/pid/emory:rm7xf; 2) Title in DiscoverE: Imagined Places: politics and narratives in a disputed indo-tibetan borderland; Legacy ETD link: https://legacy-etd.library.emory.edu/view/record/pid/emory:f3jbp. I am wondering if this has anything to do with the recent url change to Legacy ETDs that occurred in August 2017 when we deployed New ETDs?
defect
legacy etd embargo restrictions not current in discovere see examples in description sco staff noticed that revised embargo restrictions in legacy etds are not showing up in discovere two examples title in discovere protecting infants through tdap vaccination during pregnancy a qualitative analysis of the perspectives of obstetrician gynecologists legacy etd link title in discovere imagined places politics and narratives in a disputed indo tibetan borderland legacy etd link i am wondering if this has anything to do with the recent url change to legacy etds that occurred in august when we deployed new etds
1
35,485
7,753,422,122
IssuesEvent
2018-05-31 00:32:50
bridgedotnet/Bridge
https://api.github.com/repos/bridgedotnet/Bridge
closed
IList.IsFixedSize returns incorrect result
defect
`IList.IsFixedSize` returns `false` for fixed-size arrays. ### Steps To Reproduce https://deck.net/ef9cd5d649b973a9860781fe07f33252 https://dotnetfiddle.net/4QV6L1 ```csharp public class Program { public static void Main() { var arr = new int[] {1, 2, 3}; Console.WriteLine("int[].IsFixedSize = " + arr.IsFixedSize); var ilist = (IList) arr; Console.WriteLine("IList.IsFixedSize = " + ilist.IsFixedSize); } } ``` ### Expected Result Console output: ``` int[].IsFixedSize = True IList.IsFixedSize = True ``` ### Actual Result Console output: ``` int[].IsFixedSize = True IList.IsFixedSize = False ```
1.0
IList.IsFixedSize returns incorrect result - `IList.IsFixedSize` returns `false` for fixed-size arrays. ### Steps To Reproduce https://deck.net/ef9cd5d649b973a9860781fe07f33252 https://dotnetfiddle.net/4QV6L1 ```csharp public class Program { public static void Main() { var arr = new int[] {1, 2, 3}; Console.WriteLine("int[].IsFixedSize = " + arr.IsFixedSize); var ilist = (IList) arr; Console.WriteLine("IList.IsFixedSize = " + ilist.IsFixedSize); } } ``` ### Expected Result Console output: ``` int[].IsFixedSize = True IList.IsFixedSize = True ``` ### Actual Result Console output: ``` int[].IsFixedSize = True IList.IsFixedSize = False ```
defect
ilist isfixedsize returns incorrect result ilist isfixedsize returns false for fixed size arrays steps to reproduce csharp public class program public static void main var arr new int console writeline int isfixedsize arr isfixedsize var ilist ilist arr console writeline ilist isfixedsize ilist isfixedsize expected result console output int isfixedsize true ilist isfixedsize true actual result console output int isfixedsize true ilist isfixedsize false
1
288,780
21,720,716,239
IssuesEvent
2022-05-10 23:35:50
near-multicall/ui
https://api.github.com/repos/near-multicall/ui
closed
Automation Test via Airtable
bug documentation enhancement card low high dismiss / duplicate
**Discord or Telegram handle** @lennczar **Describe the bug** This is a test description **To Reproduce** 1. Test this 2. Test that 3. Automation **Expected behavior** This ends up on Github Issues **Screenshots** ![a1080.png](https://dl.airtable.com/.attachments/19da105693ddc3ea7fffc18ce2e79359/6d125262/a1080.png) **Desktop (please complete the following information):**  - OS: Ubuntu LTS 20.21  - Browser: Chrome, Firefox **Smartphone (please complete the following information):**  - Device: PC  - OS: Ubuntu LTS 20.21  - Browser: Ubuntu LTS 20.21 **Additional context** Good luck https://airtable.com/app9YvmKASJgAJaNI/tblrGw36VBqWrtwbw/recSjM6c1LQWZZF8L
1.0
Automation Test via Airtable - **Discord or Telegram handle** @lennczar **Describe the bug** This is a test description **To Reproduce** 1. Test this 2. Test that 3. Automation **Expected behavior** This ends up on Github Issues **Screenshots** ![a1080.png](https://dl.airtable.com/.attachments/19da105693ddc3ea7fffc18ce2e79359/6d125262/a1080.png) **Desktop (please complete the following information):**  - OS: Ubuntu LTS 20.21  - Browser: Chrome, Firefox **Smartphone (please complete the following information):**  - Device: PC  - OS: Ubuntu LTS 20.21  - Browser: Ubuntu LTS 20.21 **Additional context** Good luck https://airtable.com/app9YvmKASJgAJaNI/tblrGw36VBqWrtwbw/recSjM6c1LQWZZF8L
non_defect
automation test via airtable discord or telegram handle lennczar describe the bug this is a test description to reproduce test this test that automation expected behavior this ends up on github issues screenshots desktop please complete the following information   os ubuntu lts   browser chrome firefox smartphone please complete the following information   device pc   os ubuntu lts   browser ubuntu lts additional context good luck
0
481,582
13,889,348,695
IssuesEvent
2020-10-19 07:46:21
buddyboss/buddyboss-platform
https://api.github.com/repos/buddyboss/buddyboss-platform
opened
Add functionality to edit comments same as Edit Activity
feature: major priority: medium
**Is your feature request related to a problem? Please describe.** Would be great if we provide option to edit comments as well same as Edit Activty. **Describe the solution you'd like** Steps to reproduce the behavior: - Go to the activity page - check comment posted on activity **Describe alternatives you've considered** User should be able to edit his/her comments same as Facebook, Slack **Support ticket links** N/A
1.0
Add functionality to edit comments same as Edit Activity - **Is your feature request related to a problem? Please describe.** Would be great if we provide option to edit comments as well same as Edit Activty. **Describe the solution you'd like** Steps to reproduce the behavior: - Go to the activity page - check comment posted on activity **Describe alternatives you've considered** User should be able to edit his/her comments same as Facebook, Slack **Support ticket links** N/A
non_defect
add functionality to edit comments same as edit activity is your feature request related to a problem please describe would be great if we provide option to edit comments as well same as edit activty describe the solution you d like steps to reproduce the behavior go to the activity page check comment posted on activity describe alternatives you ve considered user should be able to edit his her comments same as facebook slack support ticket links n a
0
41,440
2,869,006,353
IssuesEvent
2015-06-05 22:32:01
dart-lang/sdk
https://api.github.com/repos/dart-lang/sdk
closed
add Timeout functionality for http package
Area-Pkg Pkg-Http Priority-Unassigned Triaged Type-Enhancement
*This issue was originally filed by theburnin...&#064;gmail.com* _____ **What steps will clearly show the issue / need for enhancement?** 1. when making http requests you usually want to timeout the request after a reasonable amount of time (usually in the region of seconds) 2. it would be great to have away to specify timeout as Duration when making a http request via the http/browser client/http client classes. **What version of the product are you using? On what operating system?** Dart 1.5.1, http 0.11.1+1
1.0
add Timeout functionality for http package - *This issue was originally filed by theburnin...&#064;gmail.com* _____ **What steps will clearly show the issue / need for enhancement?** 1. when making http requests you usually want to timeout the request after a reasonable amount of time (usually in the region of seconds) 2. it would be great to have away to specify timeout as Duration when making a http request via the http/browser client/http client classes. **What version of the product are you using? On what operating system?** Dart 1.5.1, http 0.11.1+1
non_defect
add timeout functionality for http package this issue was originally filed by theburnin gmail com what steps will clearly show the issue need for enhancement when making http requests you usually want to timeout the request after a reasonable amount of time usually in the region of seconds it would be great to have away to specify timeout as duration when making a http request via the http browser client http client classes what version of the product are you using on what operating system dart http
0
293,269
25,280,645,628
IssuesEvent
2022-11-16 15:33:25
vegaprotocol/frontend-monorepo
https://api.github.com/repos/vegaprotocol/frontend-monorepo
closed
Migrate explorer e2e tests to use vegawallet v2
Block Explorer Testing 🧪 chore
We need to switch to vegawallet v2, tests need to be updated and verified if no issues are caused by using the new wallet
1.0
Migrate explorer e2e tests to use vegawallet v2 - We need to switch to vegawallet v2, tests need to be updated and verified if no issues are caused by using the new wallet
non_defect
migrate explorer tests to use vegawallet we need to switch to vegawallet tests need to be updated and verified if no issues are caused by using the new wallet
0
22,741
3,690,859,910
IssuesEvent
2016-02-25 21:37:29
itmay/emu
https://api.github.com/repos/itmay/emu
closed
compilation error
auto-migrated Priority-Medium Type-Defect
``` R24 compiled in VS2008 Boost v1.47 d:\src\f1x\emu\trunk\src\gameserver\game.cpp(73) : error C2440: 'initializing' : cannot convert from 'gameServerUser_t *' to 'gameServerUser_t &' d:\src\f1x\emu\trunk\src\gameserver\game.cpp(130) : error C2440: 'initializing' : cannot convert from 'gameServerUser_t *' to 'gameServerUser_t &' d:\src\f1x\emu\trunk\src\gameserver\game.cpp(197) : error C2440: 'initializing' : cannot convert from 'gameServerUser_t *' to 'gameServerUser_t &' d:\src\f1x\emu\trunk\src\gameserver\game.cpp(241) : error C2440: 'initializing' : cannot convert from 'gameServerUser_t *' to 'gameServerUser_t &' d:\src\f1x\emu\trunk\src\gameserver\game.cpp(282) : error C2440: 'initializing' : cannot convert from 'gameServerUser_t *' to 'gameServerUser_t &' d:\src\f1x\emu\trunk\src\gameserver\game.cpp(403) : error C2102: '&' requires l-value d:\src\f1x\emu\trunk\src\gameserver\game.cpp(417) : error C2440: 'initializing' : cannot convert from 'gameServerUser_t *' to 'gameServerUser_t &' d:\src\f1x\emu\trunk\src\gameserver\game.cpp(426) : error C2039: 'count' : is not a member of 'eMUCore::socketContextManager_t' with [ T=gameServerUser_t ] d:\src\f1x\emu\trunk\src\gameserver\game.cpp(427) : error C2440: 'initializing' : cannot convert from 'gameServerUser_t *' to 'gameServerUser_t &' gameobject.cpp gameserver.cpp d:\src\f1x\emu\trunk\src\gameserver\gameserver.h(34) : error C2664: 'eMUCore::iocpEngine_t::detach' : cannot convert parameter 1 from 'gameServerUser_t' to 'eMUCore::socketContext_t *' No user-defined-conversion operator available that can perform this conversion, or the operator cannot be called d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(39) : error C2664: 'eMUCore::socketContextManager_t::socketContextManager_t(const eMUCore::socketContextManager_t &)' : cannot convert parameter 1 from 'size_t' to 'const eMUCore::socketContextManager_t &' with [ T=gameServerUser_t ] Reason: cannot convert from 'size_t' to 'const eMUCore::socketContextManager_t' with [ T=gameServerUser_t ] No constructor could take the source type, or constructor overload resolution was ambiguous d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(39) : error C2661: 'eMUCore::logger_t::logger_t' : no overloaded function takes 2 arguments d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(39) : error C2664: 'eMUCore::scheduler_t::scheduler_t(const eMUCore::scheduler_t &)' : cannot convert parameter 1 from 'eMUCore::synchronizer_t' to 'const eMUCore::scheduler_t &' Reason: cannot convert from 'eMUCore::synchronizer_t' to 'const eMUCore::scheduler_t' No user-defined-conversion operator available that can perform this conversion, or the operator cannot be called d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(39) : error C2661: 'eMUCore::iocpEngine_t::iocpEngine_t' : no overloaded function takes 2 arguments d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(39) : error C2661: 'eMUCore::tcpClient_t::tcpClient_t' : no overloaded function takes 2 arguments d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(39) : error C2661: 'eMUCore::tcpServer_t::tcpServer_t' : no overloaded function takes 4 arguments d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(39) : error C2661: 'eMUCore::udpSocket_t::udpSocket_t' : no overloaded function takes 3 arguments d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(71) : error C2660: 'eMUCore::socketContextManager_t::startup' : function does not take 3 arguments with [ T=gameServerUser_t ] d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(79) : error C2660: 'eMUCore::tcpServer_t::startup' : function does not take 0 arguments d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(83) : error C2660: 'eMUCore::udpSocket_t::startup' : function does not take 0 arguments d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(158) : error C2039: 'count' : is not a member of 'eMUCore::socketContextManager_t' with [ T=gameServerUser_t ] d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(168) : error C2039: 'count' : is not a member of 'eMUCore::socketContextManager_t' with [ T=gameServerUser_t ] d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(169) : error C2228: left of '.active' must have class/struct/union type is 'gameServerUser_t *' did you intend to use '->' instead? d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(170) : error C2664: 'gameServer_t::disconnect' : cannot convert parameter 1 from 'gameServerUser_t *' to 'gameServerUser_t &' d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(176) : error C2039: 'count' : is not a member of 'eMUCore::socketContextManager_t' with [ T=gameServerUser_t ] d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(415) : error C2664: 'eMUCore::iocpEngine_t::write' : cannot convert parameter 1 from 'gameServerUser_t' to 'eMUCore::socketContext_t *' No user-defined-conversion operator available that can perform this conversion, or the operator cannot be called d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(417) : error C2664: 'eMUCore::iocpEngine_t::write' : cannot convert parameter 1 from 'gameServerUser_t' to 'eMUCore::socketContext_t *' No user-defined-conversion operator available that can perform this conversion, or the operator cannot be called d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(428) : error C2664: 'eMUCore::iocpEngine_t::write' : cannot convert parameter 1 from 'eMUCore::tcpClient_t' to 'eMUCore::socketContext_t *' No user-defined-conversion operator available that can perform this conversion, or the operator cannot be called gate.cpp item.cpp map.cpp d:\src\f1x\emu\trunk\src\core\core.h(45) : error C3892: 'min' : you cannot assign to a variable that is const d:\src\f1x\emu\trunk\src\gameserver\map.cpp(63) : see reference to function template instantiation 'T eMUCore::role(const T &,const T &)' being compiled with [ T=size_t ] d:\src\f1x\emu\trunk\src\core\core.h(46) : error C3892: 'max' : you cannot assign to a variable that is const ``` Original issue reported on code.google.com by `felipeol...@gmail.com` on 6 Sep 2011 at 2:44 Attachments: * [BuildLog.htm](https://storage.googleapis.com/google-code-attachments/emu/issue-2/comment-0/BuildLog.htm) * [BuildLog.htm](https://storage.googleapis.com/google-code-attachments/emu/issue-2/comment-0/BuildLog.htm)
1.0
compilation error - ``` R24 compiled in VS2008 Boost v1.47 d:\src\f1x\emu\trunk\src\gameserver\game.cpp(73) : error C2440: 'initializing' : cannot convert from 'gameServerUser_t *' to 'gameServerUser_t &' d:\src\f1x\emu\trunk\src\gameserver\game.cpp(130) : error C2440: 'initializing' : cannot convert from 'gameServerUser_t *' to 'gameServerUser_t &' d:\src\f1x\emu\trunk\src\gameserver\game.cpp(197) : error C2440: 'initializing' : cannot convert from 'gameServerUser_t *' to 'gameServerUser_t &' d:\src\f1x\emu\trunk\src\gameserver\game.cpp(241) : error C2440: 'initializing' : cannot convert from 'gameServerUser_t *' to 'gameServerUser_t &' d:\src\f1x\emu\trunk\src\gameserver\game.cpp(282) : error C2440: 'initializing' : cannot convert from 'gameServerUser_t *' to 'gameServerUser_t &' d:\src\f1x\emu\trunk\src\gameserver\game.cpp(403) : error C2102: '&' requires l-value d:\src\f1x\emu\trunk\src\gameserver\game.cpp(417) : error C2440: 'initializing' : cannot convert from 'gameServerUser_t *' to 'gameServerUser_t &' d:\src\f1x\emu\trunk\src\gameserver\game.cpp(426) : error C2039: 'count' : is not a member of 'eMUCore::socketContextManager_t' with [ T=gameServerUser_t ] d:\src\f1x\emu\trunk\src\gameserver\game.cpp(427) : error C2440: 'initializing' : cannot convert from 'gameServerUser_t *' to 'gameServerUser_t &' gameobject.cpp gameserver.cpp d:\src\f1x\emu\trunk\src\gameserver\gameserver.h(34) : error C2664: 'eMUCore::iocpEngine_t::detach' : cannot convert parameter 1 from 'gameServerUser_t' to 'eMUCore::socketContext_t *' No user-defined-conversion operator available that can perform this conversion, or the operator cannot be called d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(39) : error C2664: 'eMUCore::socketContextManager_t::socketContextManager_t(const eMUCore::socketContextManager_t &)' : cannot convert parameter 1 from 'size_t' to 'const eMUCore::socketContextManager_t &' with [ T=gameServerUser_t ] Reason: cannot convert from 'size_t' to 'const eMUCore::socketContextManager_t' with [ T=gameServerUser_t ] No constructor could take the source type, or constructor overload resolution was ambiguous d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(39) : error C2661: 'eMUCore::logger_t::logger_t' : no overloaded function takes 2 arguments d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(39) : error C2664: 'eMUCore::scheduler_t::scheduler_t(const eMUCore::scheduler_t &)' : cannot convert parameter 1 from 'eMUCore::synchronizer_t' to 'const eMUCore::scheduler_t &' Reason: cannot convert from 'eMUCore::synchronizer_t' to 'const eMUCore::scheduler_t' No user-defined-conversion operator available that can perform this conversion, or the operator cannot be called d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(39) : error C2661: 'eMUCore::iocpEngine_t::iocpEngine_t' : no overloaded function takes 2 arguments d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(39) : error C2661: 'eMUCore::tcpClient_t::tcpClient_t' : no overloaded function takes 2 arguments d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(39) : error C2661: 'eMUCore::tcpServer_t::tcpServer_t' : no overloaded function takes 4 arguments d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(39) : error C2661: 'eMUCore::udpSocket_t::udpSocket_t' : no overloaded function takes 3 arguments d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(71) : error C2660: 'eMUCore::socketContextManager_t::startup' : function does not take 3 arguments with [ T=gameServerUser_t ] d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(79) : error C2660: 'eMUCore::tcpServer_t::startup' : function does not take 0 arguments d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(83) : error C2660: 'eMUCore::udpSocket_t::startup' : function does not take 0 arguments d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(158) : error C2039: 'count' : is not a member of 'eMUCore::socketContextManager_t' with [ T=gameServerUser_t ] d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(168) : error C2039: 'count' : is not a member of 'eMUCore::socketContextManager_t' with [ T=gameServerUser_t ] d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(169) : error C2228: left of '.active' must have class/struct/union type is 'gameServerUser_t *' did you intend to use '->' instead? d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(170) : error C2664: 'gameServer_t::disconnect' : cannot convert parameter 1 from 'gameServerUser_t *' to 'gameServerUser_t &' d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(176) : error C2039: 'count' : is not a member of 'eMUCore::socketContextManager_t' with [ T=gameServerUser_t ] d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(415) : error C2664: 'eMUCore::iocpEngine_t::write' : cannot convert parameter 1 from 'gameServerUser_t' to 'eMUCore::socketContext_t *' No user-defined-conversion operator available that can perform this conversion, or the operator cannot be called d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(417) : error C2664: 'eMUCore::iocpEngine_t::write' : cannot convert parameter 1 from 'gameServerUser_t' to 'eMUCore::socketContext_t *' No user-defined-conversion operator available that can perform this conversion, or the operator cannot be called d:\src\f1x\emu\trunk\src\gameserver\gameserver.cpp(428) : error C2664: 'eMUCore::iocpEngine_t::write' : cannot convert parameter 1 from 'eMUCore::tcpClient_t' to 'eMUCore::socketContext_t *' No user-defined-conversion operator available that can perform this conversion, or the operator cannot be called gate.cpp item.cpp map.cpp d:\src\f1x\emu\trunk\src\core\core.h(45) : error C3892: 'min' : you cannot assign to a variable that is const d:\src\f1x\emu\trunk\src\gameserver\map.cpp(63) : see reference to function template instantiation 'T eMUCore::role(const T &,const T &)' being compiled with [ T=size_t ] d:\src\f1x\emu\trunk\src\core\core.h(46) : error C3892: 'max' : you cannot assign to a variable that is const ``` Original issue reported on code.google.com by `felipeol...@gmail.com` on 6 Sep 2011 at 2:44 Attachments: * [BuildLog.htm](https://storage.googleapis.com/google-code-attachments/emu/issue-2/comment-0/BuildLog.htm) * [BuildLog.htm](https://storage.googleapis.com/google-code-attachments/emu/issue-2/comment-0/BuildLog.htm)
defect
compilation error compiled in boost d src emu trunk src gameserver game cpp error initializing cannot convert from gameserveruser t to gameserveruser t d src emu trunk src gameserver game cpp error initializing cannot convert from gameserveruser t to gameserveruser t d src emu trunk src gameserver game cpp error initializing cannot convert from gameserveruser t to gameserveruser t d src emu trunk src gameserver game cpp error initializing cannot convert from gameserveruser t to gameserveruser t d src emu trunk src gameserver game cpp error initializing cannot convert from gameserveruser t to gameserveruser t d src emu trunk src gameserver game cpp error requires l value d src emu trunk src gameserver game cpp error initializing cannot convert from gameserveruser t to gameserveruser t d src emu trunk src gameserver game cpp error count is not a member of emucore socketcontextmanager t with t gameserveruser t d src emu trunk src gameserver game cpp error initializing cannot convert from gameserveruser t to gameserveruser t gameobject cpp gameserver cpp d src emu trunk src gameserver gameserver h error emucore iocpengine t detach cannot convert parameter from gameserveruser t to emucore socketcontext t no user defined conversion operator available that can perform this conversion or the operator cannot be called d src emu trunk src gameserver gameserver cpp error emucore socketcontextmanager t socketcontextmanager t const emucore socketcontextmanager t cannot convert parameter from size t to const emucore socketcontextmanager t with t gameserveruser t reason cannot convert from size t to const emucore socketcontextmanager t with t gameserveruser t no constructor could take the source type or constructor overload resolution was ambiguous d src emu trunk src gameserver gameserver cpp error emucore logger t logger t no overloaded function takes arguments d src emu trunk src gameserver gameserver cpp error emucore scheduler t scheduler t const emucore scheduler t cannot convert parameter from emucore synchronizer t to const emucore scheduler t reason cannot convert from emucore synchronizer t to const emucore scheduler t no user defined conversion operator available that can perform this conversion or the operator cannot be called d src emu trunk src gameserver gameserver cpp error emucore iocpengine t iocpengine t no overloaded function takes arguments d src emu trunk src gameserver gameserver cpp error emucore tcpclient t tcpclient t no overloaded function takes arguments d src emu trunk src gameserver gameserver cpp error emucore tcpserver t tcpserver t no overloaded function takes arguments d src emu trunk src gameserver gameserver cpp error emucore udpsocket t udpsocket t no overloaded function takes arguments d src emu trunk src gameserver gameserver cpp error emucore socketcontextmanager t startup function does not take arguments with t gameserveruser t d src emu trunk src gameserver gameserver cpp error emucore tcpserver t startup function does not take arguments d src emu trunk src gameserver gameserver cpp error emucore udpsocket t startup function does not take arguments d src emu trunk src gameserver gameserver cpp error count is not a member of emucore socketcontextmanager t with t gameserveruser t d src emu trunk src gameserver gameserver cpp error count is not a member of emucore socketcontextmanager t with t gameserveruser t d src emu trunk src gameserver gameserver cpp error left of active must have class struct union type is gameserveruser t did you intend to use instead d src emu trunk src gameserver gameserver cpp error gameserver t disconnect cannot convert parameter from gameserveruser t to gameserveruser t d src emu trunk src gameserver gameserver cpp error count is not a member of emucore socketcontextmanager t with t gameserveruser t d src emu trunk src gameserver gameserver cpp error emucore iocpengine t write cannot convert parameter from gameserveruser t to emucore socketcontext t no user defined conversion operator available that can perform this conversion or the operator cannot be called d src emu trunk src gameserver gameserver cpp error emucore iocpengine t write cannot convert parameter from gameserveruser t to emucore socketcontext t no user defined conversion operator available that can perform this conversion or the operator cannot be called d src emu trunk src gameserver gameserver cpp error emucore iocpengine t write cannot convert parameter from emucore tcpclient t to emucore socketcontext t no user defined conversion operator available that can perform this conversion or the operator cannot be called gate cpp item cpp map cpp d src emu trunk src core core h error min you cannot assign to a variable that is const d src emu trunk src gameserver map cpp see reference to function template instantiation t emucore role const t const t being compiled with t size t d src emu trunk src core core h error max you cannot assign to a variable that is const original issue reported on code google com by felipeol gmail com on sep at attachments
1
38,726
12,598,101,254
IssuesEvent
2020-06-11 01:57:22
m-roll/hj
https://api.github.com/repos/m-roll/hj
closed
Do not store room code in URL slug
security
Keeping the room code in the URL slug clutters the users history with potentially expired room code URLs, and increases the odds that someone joins a room that has happens to reuse a name. Instead, invite links should be generated by including query parameters that include some kind of encoded key that includes the room code in it. The key will be an encrypted token that is unique to that instance of the room code and is kept to the server. This way, we can validate invite links and only allow the user to connect if the invite link is valid for that session.
True
Do not store room code in URL slug - Keeping the room code in the URL slug clutters the users history with potentially expired room code URLs, and increases the odds that someone joins a room that has happens to reuse a name. Instead, invite links should be generated by including query parameters that include some kind of encoded key that includes the room code in it. The key will be an encrypted token that is unique to that instance of the room code and is kept to the server. This way, we can validate invite links and only allow the user to connect if the invite link is valid for that session.
non_defect
do not store room code in url slug keeping the room code in the url slug clutters the users history with potentially expired room code urls and increases the odds that someone joins a room that has happens to reuse a name instead invite links should be generated by including query parameters that include some kind of encoded key that includes the room code in it the key will be an encrypted token that is unique to that instance of the room code and is kept to the server this way we can validate invite links and only allow the user to connect if the invite link is valid for that session
0
532,409
15,555,772,561
IssuesEvent
2021-03-16 06:44:30
wso2/product-apim
https://api.github.com/repos/wso2/product-apim
closed
[API 2.0.0 -> APIM 4.0.0 Migration] Error in oracle script
Priority/High Type/Bug migration-4.0.0 migration-4.0.0-docs
### Description: When running the oracle db script below error encountered. ALTER TABLE AM_API_COMMENTS ADD FOREIGN KEY(PARENT_COMMENT_ID) REFERENCES AM_API_COMMENTS(COMMENT_ID) ON DELETE CASCADE Error report - ORA-02270: no matching unique or primary key for this column-list 02270. 00000 - "no matching unique or primary key for this column-list" *Cause: A REFERENCES clause in a CREATE/ALTER TABLE statement gives a column-list for which there is no matching unique or primary key constraint in the referenced table. *Action: Find the correct column names using the ALL_CONS_COLUMNS catalog view
1.0
[API 2.0.0 -> APIM 4.0.0 Migration] Error in oracle script - ### Description: When running the oracle db script below error encountered. ALTER TABLE AM_API_COMMENTS ADD FOREIGN KEY(PARENT_COMMENT_ID) REFERENCES AM_API_COMMENTS(COMMENT_ID) ON DELETE CASCADE Error report - ORA-02270: no matching unique or primary key for this column-list 02270. 00000 - "no matching unique or primary key for this column-list" *Cause: A REFERENCES clause in a CREATE/ALTER TABLE statement gives a column-list for which there is no matching unique or primary key constraint in the referenced table. *Action: Find the correct column names using the ALL_CONS_COLUMNS catalog view
non_defect
error in oracle script description when running the oracle db script below error encountered alter table am api comments add foreign key parent comment id references am api comments comment id on delete cascade error report ora no matching unique or primary key for this column list no matching unique or primary key for this column list cause a references clause in a create alter table statement gives a column list for which there is no matching unique or primary key constraint in the referenced table action find the correct column names using the all cons columns catalog view
0
113,397
17,139,734,340
IssuesEvent
2021-07-13 08:18:16
tabac-ws-demo/java-demo
https://api.github.com/repos/tabac-ws-demo/java-demo
opened
CVE-2020-2933 (Low) detected in mysql-connector-java-5.1.25.jar
security vulnerability
## CVE-2020-2933 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mysql-connector-java-5.1.25.jar</b></p></summary> <p>MySQL JDBC Type 4 driver</p> <p>Library home page: <a href="http://dev.mysql.com/doc/connector-j/en/">http://dev.mysql.com/doc/connector-j/en/</a></p> <p>Path to dependency file: java-demo/pom.xml</p> <p>Path to vulnerable library: canner/.m2/repository/mysql/mysql-connector-java/5.1.25/mysql-connector-java-5.1.25.jar,java-demo/target/easybuggy-1-SNAPSHOT/WEB-INF/lib/mysql-connector-java-5.1.25.jar</p> <p> Dependency Hierarchy: - :x: **mysql-connector-java-5.1.25.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/tabac-ws-demo/java-demo/commit/bd54e7a149d715d50db77eb4a258837ca6b89d73">bd54e7a149d715d50db77eb4a258837ca6b89d73</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Vulnerability in the MySQL Connectors product of Oracle MySQL (component: Connector/J). Supported versions that are affected are 5.1.48 and prior. Difficult to exploit vulnerability allows high privileged attacker with network access via multiple protocols to compromise MySQL Connectors. Successful attacks of this vulnerability can result in unauthorized ability to cause a partial denial of service (partial DOS) of MySQL Connectors. CVSS 3.0 Base Score 2.2 (Availability impacts). CVSS Vector: (CVSS:3.0/AV:N/AC:H/PR:H/UI:N/S:U/C:N/I:N/A:L). <p>Publish Date: 2020-04-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-2933>CVE-2020-2933</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>2.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://docs.oracle.com/javase/7/docs/api/javax/xml/XMLConstants.html#FEATURE_SECURE_PROCESSING">https://docs.oracle.com/javase/7/docs/api/javax/xml/XMLConstants.html#FEATURE_SECURE_PROCESSING</a></p> <p>Release Date: 2020-04-15</p> <p>Fix Resolution: mysql:mysql-connector-java:5.1.49</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"mysql","packageName":"mysql-connector-java","packageVersion":"5.1.25","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"mysql:mysql-connector-java:5.1.25","isMinimumFixVersionAvailable":true,"minimumFixVersion":"mysql:mysql-connector-java:5.1.49"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2020-2933","vulnerabilityDetails":"Vulnerability in the MySQL Connectors product of Oracle MySQL (component: Connector/J). Supported versions that are affected are 5.1.48 and prior. Difficult to exploit vulnerability allows high privileged attacker with network access via multiple protocols to compromise MySQL Connectors. Successful attacks of this vulnerability can result in unauthorized ability to cause a partial denial of service (partial DOS) of MySQL Connectors. CVSS 3.0 Base Score 2.2 (Availability impacts). CVSS Vector: (CVSS:3.0/AV:N/AC:H/PR:H/UI:N/S:U/C:N/I:N/A:L).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-2933","cvss3Severity":"low","cvss3Score":"2.2","cvss3Metrics":{"A":"Low","AC":"High","PR":"High","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-2933 (Low) detected in mysql-connector-java-5.1.25.jar - ## CVE-2020-2933 - Low Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mysql-connector-java-5.1.25.jar</b></p></summary> <p>MySQL JDBC Type 4 driver</p> <p>Library home page: <a href="http://dev.mysql.com/doc/connector-j/en/">http://dev.mysql.com/doc/connector-j/en/</a></p> <p>Path to dependency file: java-demo/pom.xml</p> <p>Path to vulnerable library: canner/.m2/repository/mysql/mysql-connector-java/5.1.25/mysql-connector-java-5.1.25.jar,java-demo/target/easybuggy-1-SNAPSHOT/WEB-INF/lib/mysql-connector-java-5.1.25.jar</p> <p> Dependency Hierarchy: - :x: **mysql-connector-java-5.1.25.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/tabac-ws-demo/java-demo/commit/bd54e7a149d715d50db77eb4a258837ca6b89d73">bd54e7a149d715d50db77eb4a258837ca6b89d73</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Vulnerability in the MySQL Connectors product of Oracle MySQL (component: Connector/J). Supported versions that are affected are 5.1.48 and prior. Difficult to exploit vulnerability allows high privileged attacker with network access via multiple protocols to compromise MySQL Connectors. Successful attacks of this vulnerability can result in unauthorized ability to cause a partial denial of service (partial DOS) of MySQL Connectors. CVSS 3.0 Base Score 2.2 (Availability impacts). CVSS Vector: (CVSS:3.0/AV:N/AC:H/PR:H/UI:N/S:U/C:N/I:N/A:L). <p>Publish Date: 2020-04-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-2933>CVE-2020-2933</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>2.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://docs.oracle.com/javase/7/docs/api/javax/xml/XMLConstants.html#FEATURE_SECURE_PROCESSING">https://docs.oracle.com/javase/7/docs/api/javax/xml/XMLConstants.html#FEATURE_SECURE_PROCESSING</a></p> <p>Release Date: 2020-04-15</p> <p>Fix Resolution: mysql:mysql-connector-java:5.1.49</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"mysql","packageName":"mysql-connector-java","packageVersion":"5.1.25","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"mysql:mysql-connector-java:5.1.25","isMinimumFixVersionAvailable":true,"minimumFixVersion":"mysql:mysql-connector-java:5.1.49"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2020-2933","vulnerabilityDetails":"Vulnerability in the MySQL Connectors product of Oracle MySQL (component: Connector/J). Supported versions that are affected are 5.1.48 and prior. Difficult to exploit vulnerability allows high privileged attacker with network access via multiple protocols to compromise MySQL Connectors. Successful attacks of this vulnerability can result in unauthorized ability to cause a partial denial of service (partial DOS) of MySQL Connectors. CVSS 3.0 Base Score 2.2 (Availability impacts). CVSS Vector: (CVSS:3.0/AV:N/AC:H/PR:H/UI:N/S:U/C:N/I:N/A:L).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-2933","cvss3Severity":"low","cvss3Score":"2.2","cvss3Metrics":{"A":"Low","AC":"High","PR":"High","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_defect
cve low detected in mysql connector java jar cve low severity vulnerability vulnerable library mysql connector java jar mysql jdbc type driver library home page a href path to dependency file java demo pom xml path to vulnerable library canner repository mysql mysql connector java mysql connector java jar java demo target easybuggy snapshot web inf lib mysql connector java jar dependency hierarchy x mysql connector java jar vulnerable library found in head commit a href found in base branch main vulnerability details vulnerability in the mysql connectors product of oracle mysql component connector j supported versions that are affected are and prior difficult to exploit vulnerability allows high privileged attacker with network access via multiple protocols to compromise mysql connectors successful attacks of this vulnerability can result in unauthorized ability to cause a partial denial of service partial dos of mysql connectors cvss base score availability impacts cvss vector cvss av n ac h pr h ui n s u c n i n a l publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required high user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution mysql mysql connector java rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree mysql mysql connector java isminimumfixversionavailable true minimumfixversion mysql mysql connector java basebranches vulnerabilityidentifier cve vulnerabilitydetails vulnerability in the mysql connectors product of oracle mysql component connector j supported versions that are affected are and prior difficult to exploit vulnerability allows high privileged attacker with network access via multiple protocols to compromise mysql connectors successful attacks of this vulnerability can result in unauthorized ability to cause a partial denial of service partial dos of mysql connectors cvss base score availability impacts cvss vector cvss av n ac h pr h ui n s u c n i n a l vulnerabilityurl
0
674,917
23,070,363,810
IssuesEvent
2022-07-25 17:26:28
googleapis/repo-automation-bots
https://api.github.com/repos/googleapis/repo-automation-bots
closed
[sync-repo-settings]: support "requiresLinearHistory" in branchProtectionRule
type: feature request priority: p3
Branch protection rules allow maintainers to require a linear commit history, or not. Requiring it disallows the use of Merge Commits. While Merge Commits are generally not encouraged in `googleapis` repos, they are idiomatic means of keeping a branch in sync with another. So if we are working a feature branch that we want to stay in sync with `main`, we must use a Merge commit with the changes from `main` into a feature branch. By default `require linear history` is enabled, which disallows merge commits on the branch. I'd like to be able to disable it on a per-branch basis. ![image](https://user-images.githubusercontent.com/6644735/180332173-fa06db46-47bc-48f6-8d2b-d7c655b739d1.png)
1.0
[sync-repo-settings]: support "requiresLinearHistory" in branchProtectionRule - Branch protection rules allow maintainers to require a linear commit history, or not. Requiring it disallows the use of Merge Commits. While Merge Commits are generally not encouraged in `googleapis` repos, they are idiomatic means of keeping a branch in sync with another. So if we are working a feature branch that we want to stay in sync with `main`, we must use a Merge commit with the changes from `main` into a feature branch. By default `require linear history` is enabled, which disallows merge commits on the branch. I'd like to be able to disable it on a per-branch basis. ![image](https://user-images.githubusercontent.com/6644735/180332173-fa06db46-47bc-48f6-8d2b-d7c655b739d1.png)
non_defect
support requireslinearhistory in branchprotectionrule branch protection rules allow maintainers to require a linear commit history or not requiring it disallows the use of merge commits while merge commits are generally not encouraged in googleapis repos they are idiomatic means of keeping a branch in sync with another so if we are working a feature branch that we want to stay in sync with main we must use a merge commit with the changes from main into a feature branch by default require linear history is enabled which disallows merge commits on the branch i d like to be able to disable it on a per branch basis
0
341,106
24,683,139,271
IssuesEvent
2022-10-18 23:59:35
smpawar/testrepo2
https://api.github.com/repos/smpawar/testrepo2
opened
DOC: migration doc throws error 404
documentation
### Documentation change request Details? Change the url ### Version Latest ### What database engine are you seeing the problem on? Oracle ### Relevant log output ```shell fix the issue ```
1.0
DOC: migration doc throws error 404 - ### Documentation change request Details? Change the url ### Version Latest ### What database engine are you seeing the problem on? Oracle ### Relevant log output ```shell fix the issue ```
non_defect
doc migration doc throws error documentation change request details change the url version latest what database engine are you seeing the problem on oracle relevant log output shell fix the issue
0
9,287
2,615,143,121
IssuesEvent
2015-03-01 06:18:23
chrsmith/html5rocks
https://api.github.com/repos/chrsmith/html5rocks
closed
Crashes in top of tree Chrome
auto-migrated Priority-Medium Type-Defect
``` Please describe the issue: Get the top of tree Chrome. Go to www.html5rocks.com Get an "Aw, Snap!" screen. Please provide any additional information below. ``` Original issue reported on code.google.com by `g...@google.com` on 22 Jun 2010 at 11:12
1.0
Crashes in top of tree Chrome - ``` Please describe the issue: Get the top of tree Chrome. Go to www.html5rocks.com Get an "Aw, Snap!" screen. Please provide any additional information below. ``` Original issue reported on code.google.com by `g...@google.com` on 22 Jun 2010 at 11:12
defect
crashes in top of tree chrome please describe the issue get the top of tree chrome go to get an aw snap screen please provide any additional information below original issue reported on code google com by g google com on jun at
1
209,186
16,177,812,287
IssuesEvent
2021-05-03 09:48:45
Arquisoft/radarin_es2b
https://api.github.com/repos/Arquisoft/radarin_es2b
closed
Actualizar la documentación
documentation enhancement
Hay que terminar de actualizar la documentación, puesto que se encuentra en un estado que se corresponde con fases anteriores del proyecto y no con la actual.
1.0
Actualizar la documentación - Hay que terminar de actualizar la documentación, puesto que se encuentra en un estado que se corresponde con fases anteriores del proyecto y no con la actual.
non_defect
actualizar la documentación hay que terminar de actualizar la documentación puesto que se encuentra en un estado que se corresponde con fases anteriores del proyecto y no con la actual
0
701,325
24,095,360,265
IssuesEvent
2022-09-19 18:14:21
darktable-org/darktable
https://api.github.com/repos/darktable-org/darktable
closed
darktable current master (Mac OSX build) crashes when opening an image in darkroom. White balance issue ?
priority: high understood: unclear bug: pending
**Describe the bug/issue** darktable opens and import an image (.nef file from Nikon) in lighttable and crashes when trying to open the image in darkroom. lldb suggests an issue with white balance (temperature.c) I have been using current master OSX builds from MStraeten The last build that works fine is: [darktable-4.1.0+236~g78747838c_x86_64.dmg](https://mega.nz/file/OWIGXI5Y#xcO_8v8bPSkMEuhX62j06j56ztSKLPDhFYMlXYoal7o) The next one is the first that showed the issue: [darktable-4.1.0+283~g5cca9d64b_x86_64.dmg](https://mega.nz/file/XPQWHB7I#ecADJh48CM9VDQgqfp6azLRMy9DtUXnLQhuYTCnLaX4) And the following ones all showed the issue up to the last one Martin build: [darktable-4.1.0+385~gf2ebe6b26_x86_64.dmg](https://mega.nz/file/vDgkDB6Q#8QUlYYscPPGa_rVyDUHfP8olN1l6zlVSsLgfUYEe_Rs) **To Reproduce** Clean install darktable Import an image with no previous .xmp Open the image in darkroom darktable crashes **Expected behavior** Open the image in darkroom **Screenshots** _(if applicable)_ **Screencast** _(if applicable)_ **Which commit introduced the error** The last build that works fine is: [darktable-4.1.0+236~g78747838c_x86_64.dmg](https://mega.nz/file/OWIGXI5Y#xcO_8v8bPSkMEuhX62j06j56ztSKLPDhFYMlXYoal7o) The next one is the first that showed the issue: [darktable-4.1.0+283~g5cca9d64b_x86_64.dmg](https://mega.nz/file/XPQWHB7I#ecADJh48CM9VDQgqfp6azLRMy9DtUXnLQhuYTCnLaX4) And the following ones all showed the issue up to the last one Martin build: [darktable-4.1.0+385~gf2ebe6b26_x86_64.dmg](https://mega.nz/file/vDgkDB6Q#8QUlYYscPPGa_rVyDUHfP8olN1l6zlVSsLgfUYEe_Rs) _A bisect is much appreciated and can significantly simplify the developer's job._ I run lldb, see attached file, and it suggests (actually Martin Straeten review the file and suggested) an issue with white balance (temperature.c). **Platform** _Please fill as much information as possible in the list given below. Please state "unknown" where you do not know the answer and remove any sections that are not applicable _ * darktable version : darktable-4.1.0+283 and following builds * OS : MacOS Monterey 12.6 * Memory : 64 GB * Graphics card : AMD Radeon Pro 5700 xt 16 GB * Graphics driver : * OpenCL installed : yes * OpenCL activated : yes * Xorg : * Desktop : * GTK+ : * gcc : * cflags : * CMAKE_BUILD_TYPE : [Mas OSX darktable issue lldb results.txt](https://github.com/darktable-org/darktable/files/9597341/Mas.OSX.darktable.issue.lldb.results.txt) **Additional context** _Please provide any additional information you think may be useful, for example:_ - Can you reproduce with another darktable version(s)? yes , see above - Can you reproduce with a RAW or Jpeg or both? Issue with RAW, JPG opens fine (limited testing) - Are the steps above reproducible with a fresh edit (i.e. after discarding history)? yes - If the issue is with the output image, attach an XMP file : file attached of an image that failed to open in darkroom: DSE_1016.NEF.txt - Is the issue still present using an empty/new config-dir (e.g. start darktable with --configdir "/tmp")? yes - Do you use lua scripts? No - What lua scripts start automatically? - What lua scripts were running when the bug occurred? [Mas OSX darktable issue lldb results.txt](https://github.com/darktable-org/darktable/files/9597344/Mas.OSX.darktable.issue.lldb.results.txt) [DSE_1016.NEF.txt](https://github.com/darktable-org/darktable/files/9597490/DSE_1016.NEF.txt)
1.0
darktable current master (Mac OSX build) crashes when opening an image in darkroom. White balance issue ? - **Describe the bug/issue** darktable opens and import an image (.nef file from Nikon) in lighttable and crashes when trying to open the image in darkroom. lldb suggests an issue with white balance (temperature.c) I have been using current master OSX builds from MStraeten The last build that works fine is: [darktable-4.1.0+236~g78747838c_x86_64.dmg](https://mega.nz/file/OWIGXI5Y#xcO_8v8bPSkMEuhX62j06j56ztSKLPDhFYMlXYoal7o) The next one is the first that showed the issue: [darktable-4.1.0+283~g5cca9d64b_x86_64.dmg](https://mega.nz/file/XPQWHB7I#ecADJh48CM9VDQgqfp6azLRMy9DtUXnLQhuYTCnLaX4) And the following ones all showed the issue up to the last one Martin build: [darktable-4.1.0+385~gf2ebe6b26_x86_64.dmg](https://mega.nz/file/vDgkDB6Q#8QUlYYscPPGa_rVyDUHfP8olN1l6zlVSsLgfUYEe_Rs) **To Reproduce** Clean install darktable Import an image with no previous .xmp Open the image in darkroom darktable crashes **Expected behavior** Open the image in darkroom **Screenshots** _(if applicable)_ **Screencast** _(if applicable)_ **Which commit introduced the error** The last build that works fine is: [darktable-4.1.0+236~g78747838c_x86_64.dmg](https://mega.nz/file/OWIGXI5Y#xcO_8v8bPSkMEuhX62j06j56ztSKLPDhFYMlXYoal7o) The next one is the first that showed the issue: [darktable-4.1.0+283~g5cca9d64b_x86_64.dmg](https://mega.nz/file/XPQWHB7I#ecADJh48CM9VDQgqfp6azLRMy9DtUXnLQhuYTCnLaX4) And the following ones all showed the issue up to the last one Martin build: [darktable-4.1.0+385~gf2ebe6b26_x86_64.dmg](https://mega.nz/file/vDgkDB6Q#8QUlYYscPPGa_rVyDUHfP8olN1l6zlVSsLgfUYEe_Rs) _A bisect is much appreciated and can significantly simplify the developer's job._ I run lldb, see attached file, and it suggests (actually Martin Straeten review the file and suggested) an issue with white balance (temperature.c). **Platform** _Please fill as much information as possible in the list given below. Please state "unknown" where you do not know the answer and remove any sections that are not applicable _ * darktable version : darktable-4.1.0+283 and following builds * OS : MacOS Monterey 12.6 * Memory : 64 GB * Graphics card : AMD Radeon Pro 5700 xt 16 GB * Graphics driver : * OpenCL installed : yes * OpenCL activated : yes * Xorg : * Desktop : * GTK+ : * gcc : * cflags : * CMAKE_BUILD_TYPE : [Mas OSX darktable issue lldb results.txt](https://github.com/darktable-org/darktable/files/9597341/Mas.OSX.darktable.issue.lldb.results.txt) **Additional context** _Please provide any additional information you think may be useful, for example:_ - Can you reproduce with another darktable version(s)? yes , see above - Can you reproduce with a RAW or Jpeg or both? Issue with RAW, JPG opens fine (limited testing) - Are the steps above reproducible with a fresh edit (i.e. after discarding history)? yes - If the issue is with the output image, attach an XMP file : file attached of an image that failed to open in darkroom: DSE_1016.NEF.txt - Is the issue still present using an empty/new config-dir (e.g. start darktable with --configdir "/tmp")? yes - Do you use lua scripts? No - What lua scripts start automatically? - What lua scripts were running when the bug occurred? [Mas OSX darktable issue lldb results.txt](https://github.com/darktable-org/darktable/files/9597344/Mas.OSX.darktable.issue.lldb.results.txt) [DSE_1016.NEF.txt](https://github.com/darktable-org/darktable/files/9597490/DSE_1016.NEF.txt)
non_defect
darktable current master mac osx build crashes when opening an image in darkroom white balance issue describe the bug issue darktable opens and import an image nef file from nikon in lighttable and crashes when trying to open the image in darkroom lldb suggests an issue with white balance temperature c i have been using current master osx builds from mstraeten the last build that works fine is the next one is the first that showed the issue and the following ones all showed the issue up to the last one martin build to reproduce clean install darktable import an image with no previous xmp open the image in darkroom darktable crashes expected behavior open the image in darkroom screenshots if applicable screencast if applicable which commit introduced the error the last build that works fine is the next one is the first that showed the issue and the following ones all showed the issue up to the last one martin build a bisect is much appreciated and can significantly simplify the developer s job i run lldb see attached file and it suggests actually martin straeten review the file and suggested an issue with white balance temperature c platform please fill as much information as possible in the list given below please state unknown where you do not know the answer and remove any sections that are not applicable darktable version darktable and following builds os macos monterey memory gb graphics card amd radeon pro xt gb graphics driver opencl installed yes opencl activated yes xorg desktop gtk gcc cflags cmake build type additional context please provide any additional information you think may be useful for example can you reproduce with another darktable version s yes see above can you reproduce with a raw or jpeg or both issue with raw jpg opens fine limited testing are the steps above reproducible with a fresh edit i e after discarding history yes if the issue is with the output image attach an xmp file file attached of an image that failed to open in darkroom dse nef txt is the issue still present using an empty new config dir e g start darktable with configdir tmp yes do you use lua scripts no what lua scripts start automatically what lua scripts were running when the bug occurred
0
28,616
12,891,605,726
IssuesEvent
2020-07-13 18:02:10
thkl/hap-homematic
https://api.github.com/repos/thkl/hap-homematic
opened
Absturz nach Löschen von HVL
DeviceService enhancement
Hallo meine HAP Installation ist leider komplett abgeschmiert. Auch nach mehrmaligem Neuinstallieren, bekomme ich es nicht wieder zum Laufen. Ich habe HVL (Homematic Virtual Interface) deinstalliert. https://github.com/thkl/Homematic-Virtual-Interface Nach dieser Anleitung: > How to remove the stuff: > There is currently no automated process so you have to do it step by step. 1st make sure the hvl is still running (ccu is not able to remove objects without response from interface) 2nd remove all your virtual devices from your ccu's device list 3rd remove the changes in in /etc/config_templates/InterfacesList.xml 4rd run this to remove the interface from regadom at your ccu's webinterface (Programs/Test Script): dom.DeleteObject(dom.GetObject('HVL')); 5th reboot your ccu Jetzt geht nichts mehr. Ich vermute, es ist irgendetwas von diesem Interface übrig geblieben. HAP hängt sich beim Starten, beim Einlesen von Interfaces auf. Hier der letzte Eintrag in /var/log/hap-homematic.log > [Mon Jul 13 2020 19:58:37 GMT+0200 (CEST)] debug - [HAP Server] [Rega] RegaScript string sifId;boolean df = true;Write('{"interfaces":[');foreach(sifId, root.Interfaces().EnumIDs()){object oIf = dom.GetObject(sifId);if(df) {df = false;} else { Write(',');}Write('{')Write('"id": ' # sifId # ',');Write('"name": "' # oIf.Name() # '",');Write('"type": "' # oIf.Type() # '",');Write('"typename": "' # oIf.TypeName() # '",');Write('"info": "' # oIf.InterfaceInfo() # '",');Write('"url": "' # oIf.InterfaceUrl() # '"');Write('}');} Write(']}'); > [Mon Jul 13 2020 19:58:37 GMT+0200 (CEST)] debug - [HAP Server] [Rega] result is {"interfaces":[{"id": 1009,"name": "BidCos-RF","type": "458753","typename": "INTERFACE","info": "BidCos-RF","url": "xmlrpc_bin://127.0.0.1:32001"},{"id": 1010,"name": "VirtualDevices","type": "458753","typename": "INTERFACE","info": "Virtual Devices","url": "xmlrpc://127.0.0.1:39292/groups"},{"id": 1011,"name": "HmIP-RF","type": "458753","typename": "INTERFACE","info": "HmIP-RF","url": "xmlrpc://127.0.0.1:32010"},{"id": 1503,"name": "CUxD","type": "458753","typename": "INTERFACE","info": "CUxD","url": "xmlrpc_bin://127.0.0.1:8701"},{"id": 1504,"name": " Kann jemand helfen, wie es wieder zum Laufen bekomme? Die Webseite ist zwar erreichbar, aber man kann keine Kofiguration anlegen. Auch eine vorherige Sicherung einspielen bringt keinen Erfolg. Vielen Dank Christian
1.0
Absturz nach Löschen von HVL - Hallo meine HAP Installation ist leider komplett abgeschmiert. Auch nach mehrmaligem Neuinstallieren, bekomme ich es nicht wieder zum Laufen. Ich habe HVL (Homematic Virtual Interface) deinstalliert. https://github.com/thkl/Homematic-Virtual-Interface Nach dieser Anleitung: > How to remove the stuff: > There is currently no automated process so you have to do it step by step. 1st make sure the hvl is still running (ccu is not able to remove objects without response from interface) 2nd remove all your virtual devices from your ccu's device list 3rd remove the changes in in /etc/config_templates/InterfacesList.xml 4rd run this to remove the interface from regadom at your ccu's webinterface (Programs/Test Script): dom.DeleteObject(dom.GetObject('HVL')); 5th reboot your ccu Jetzt geht nichts mehr. Ich vermute, es ist irgendetwas von diesem Interface übrig geblieben. HAP hängt sich beim Starten, beim Einlesen von Interfaces auf. Hier der letzte Eintrag in /var/log/hap-homematic.log > [Mon Jul 13 2020 19:58:37 GMT+0200 (CEST)] debug - [HAP Server] [Rega] RegaScript string sifId;boolean df = true;Write('{"interfaces":[');foreach(sifId, root.Interfaces().EnumIDs()){object oIf = dom.GetObject(sifId);if(df) {df = false;} else { Write(',');}Write('{')Write('"id": ' # sifId # ',');Write('"name": "' # oIf.Name() # '",');Write('"type": "' # oIf.Type() # '",');Write('"typename": "' # oIf.TypeName() # '",');Write('"info": "' # oIf.InterfaceInfo() # '",');Write('"url": "' # oIf.InterfaceUrl() # '"');Write('}');} Write(']}'); > [Mon Jul 13 2020 19:58:37 GMT+0200 (CEST)] debug - [HAP Server] [Rega] result is {"interfaces":[{"id": 1009,"name": "BidCos-RF","type": "458753","typename": "INTERFACE","info": "BidCos-RF","url": "xmlrpc_bin://127.0.0.1:32001"},{"id": 1010,"name": "VirtualDevices","type": "458753","typename": "INTERFACE","info": "Virtual Devices","url": "xmlrpc://127.0.0.1:39292/groups"},{"id": 1011,"name": "HmIP-RF","type": "458753","typename": "INTERFACE","info": "HmIP-RF","url": "xmlrpc://127.0.0.1:32010"},{"id": 1503,"name": "CUxD","type": "458753","typename": "INTERFACE","info": "CUxD","url": "xmlrpc_bin://127.0.0.1:8701"},{"id": 1504,"name": " Kann jemand helfen, wie es wieder zum Laufen bekomme? Die Webseite ist zwar erreichbar, aber man kann keine Kofiguration anlegen. Auch eine vorherige Sicherung einspielen bringt keinen Erfolg. Vielen Dank Christian
non_defect
absturz nach löschen von hvl hallo meine hap installation ist leider komplett abgeschmiert auch nach mehrmaligem neuinstallieren bekomme ich es nicht wieder zum laufen ich habe hvl homematic virtual interface deinstalliert nach dieser anleitung how to remove the stuff there is currently no automated process so you have to do it step by step make sure the hvl is still running ccu is not able to remove objects without response from interface remove all your virtual devices from your ccu s device list remove the changes in in etc config templates interfaceslist xml run this to remove the interface from regadom at your ccu s webinterface programs test script dom deleteobject dom getobject hvl reboot your ccu jetzt geht nichts mehr ich vermute es ist irgendetwas von diesem interface übrig geblieben hap hängt sich beim starten beim einlesen von interfaces auf hier der letzte eintrag in var log hap homematic log debug regascript string sifid boolean df true write interfaces debug result is interfaces id name bidcos rf type typename interface info bidcos rf url xmlrpc bin id name virtualdevices type typename interface info virtual devices url xmlrpc groups id name hmip rf type typename interface info hmip rf url xmlrpc id name cuxd type typename interface info cuxd url xmlrpc bin id name kann jemand helfen wie es wieder zum laufen bekomme die webseite ist zwar erreichbar aber man kann keine kofiguration anlegen auch eine vorherige sicherung einspielen bringt keinen erfolg vielen dank christian
0
70,420
23,158,843,983
IssuesEvent
2022-07-29 15:28:49
E1337Kat/cyberpunk2077_ext_redux
https://api.github.com/repos/E1337Kat/cyberpunk2077_ext_redux
closed
Redscript Basedir Layout Hits Both Basedir And Canonical In MultiType
defect Mod
Redscript canonical layout doesn’t check whether there's files in the basedir because installers.redscript handles this by picking only one of them. MultiType will match both. Ultimately we’d probably want to move to a proper `File` type hierarchy that would encode modifications and thus prevent multiple layouts from using the same one. In the meanwhile, it's either adding a check to Canon layout to exclude basedir, or probably the better solution of creating a similar 'allowed in other types' instruction generator that’s already set up for a couple other types.
1.0
Redscript Basedir Layout Hits Both Basedir And Canonical In MultiType - Redscript canonical layout doesn’t check whether there's files in the basedir because installers.redscript handles this by picking only one of them. MultiType will match both. Ultimately we’d probably want to move to a proper `File` type hierarchy that would encode modifications and thus prevent multiple layouts from using the same one. In the meanwhile, it's either adding a check to Canon layout to exclude basedir, or probably the better solution of creating a similar 'allowed in other types' instruction generator that’s already set up for a couple other types.
defect
redscript basedir layout hits both basedir and canonical in multitype redscript canonical layout doesn’t check whether there s files in the basedir because installers redscript handles this by picking only one of them multitype will match both ultimately we’d probably want to move to a proper file type hierarchy that would encode modifications and thus prevent multiple layouts from using the same one in the meanwhile it s either adding a check to canon layout to exclude basedir or probably the better solution of creating a similar allowed in other types instruction generator that’s already set up for a couple other types
1
21,917
14,934,584,120
IssuesEvent
2021-01-25 10:43:46
PostHog/posthog
https://api.github.com/repos/PostHog/posthog
opened
Large events get ignored by kafka
bug clickhouse infrastructure
## Bug description See https://github.com/PostHog/posthog/issues/2927. Large kafka messages (>1MB) get just ignored by kafka. This causes problems for session recording, but probably causes other data to go missing as well. This becomes even more of an issue as we're switching to using plugins-based ingestion as everything is reading from WAL. ## Expected behavior Events with >1MB payloads do not just go missing - we return a sensible error message to client. ## How to reproduce ## Environment - [x] PostHog Cloud - [ ] self-hosted PostHog, version/commit: _please provide_ ## Additional context We're planning on switching to hosting kafka ourselves on AWS which should allow us to update the setting. cc @fuziontech @Twixes @mariusandra for pluginsland #### *Thank you* for your bug report – we love squashing them!
1.0
Large events get ignored by kafka - ## Bug description See https://github.com/PostHog/posthog/issues/2927. Large kafka messages (>1MB) get just ignored by kafka. This causes problems for session recording, but probably causes other data to go missing as well. This becomes even more of an issue as we're switching to using plugins-based ingestion as everything is reading from WAL. ## Expected behavior Events with >1MB payloads do not just go missing - we return a sensible error message to client. ## How to reproduce ## Environment - [x] PostHog Cloud - [ ] self-hosted PostHog, version/commit: _please provide_ ## Additional context We're planning on switching to hosting kafka ourselves on AWS which should allow us to update the setting. cc @fuziontech @Twixes @mariusandra for pluginsland #### *Thank you* for your bug report – we love squashing them!
non_defect
large events get ignored by kafka bug description see large kafka messages get just ignored by kafka this causes problems for session recording but probably causes other data to go missing as well this becomes even more of an issue as we re switching to using plugins based ingestion as everything is reading from wal expected behavior events with payloads do not just go missing we return a sensible error message to client how to reproduce environment posthog cloud self hosted posthog version commit please provide additional context we re planning on switching to hosting kafka ourselves on aws which should allow us to update the setting cc fuziontech twixes mariusandra for pluginsland thank you for your bug report – we love squashing them
0
734,084
25,337,760,131
IssuesEvent
2022-11-18 18:23:36
ramp4-pcar4/ramp4-pcar4
https://api.github.com/repos/ramp4-pcar4/ramp4-pcar4
closed
Yet another legend visibility bug
effort: small flavour: bug priority: must type: corrective
When an exclusive set is turned off and the last visible item has the visibility control disabled (but it was turned off by using the override property), turning the exclusive set back on will show the parent's checkbox as checked but none of the children will turn on (since the last visible one has vis control disabled). Will add a gif once I figure out how (GitHub not letting me atm). A solution would be to find another valid child to turn on if the last visible one fails.
1.0
Yet another legend visibility bug - When an exclusive set is turned off and the last visible item has the visibility control disabled (but it was turned off by using the override property), turning the exclusive set back on will show the parent's checkbox as checked but none of the children will turn on (since the last visible one has vis control disabled). Will add a gif once I figure out how (GitHub not letting me atm). A solution would be to find another valid child to turn on if the last visible one fails.
non_defect
yet another legend visibility bug when an exclusive set is turned off and the last visible item has the visibility control disabled but it was turned off by using the override property turning the exclusive set back on will show the parent s checkbox as checked but none of the children will turn on since the last visible one has vis control disabled will add a gif once i figure out how github not letting me atm a solution would be to find another valid child to turn on if the last visible one fails
0
33,727
7,204,393,587
IssuesEvent
2018-02-06 12:33:11
primefaces/primereact
https://api.github.com/repos/primefaces/primereact
closed
Duplicate identifier 'any': PickList.d.ts
defect
Error message: /node_modules/primereact/components/picklist/PickList.d.ts [INFO] (17,38): error TS2300: Duplicate identifier 'any'. The problem seems to be in this line: onChange?({event: Event, source: any, target: any}): void; And it can be solved like this: onChange?(i: {event: Event, source: any, target: any}): void;
1.0
Duplicate identifier 'any': PickList.d.ts - Error message: /node_modules/primereact/components/picklist/PickList.d.ts [INFO] (17,38): error TS2300: Duplicate identifier 'any'. The problem seems to be in this line: onChange?({event: Event, source: any, target: any}): void; And it can be solved like this: onChange?(i: {event: Event, source: any, target: any}): void;
defect
duplicate identifier any picklist d ts error message node modules primereact components picklist picklist d ts error duplicate identifier any the problem seems to be in this line onchange event event source any target any void and it can be solved like this onchange i event event source any target any void
1
50,197
13,187,375,473
IssuesEvent
2020-08-13 03:12:55
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
closed
a_nocompression.py test fails badly in DOMcalibrator. (Trac #294)
Migrated from Trac combo core defect
The WFs produced are complete craps, loads of NAN values, etc. This has been broken for a while now and gone unnoticed. The data used is ancient (IC9) and maybe DOMcalibrator no longer is happy with this. Update to newer data. DOMcalibrator is on the decline anyway. <details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/294 , reported by blaufuss and owned by blaufuss_</summary> <p> ```json { "status": "closed", "changetime": "2014-11-23T03:37:57", "description": "The WFs produced are complete craps, loads of NAN values, etc. This\nhas been broken for a while now and gone unnoticed. The\ndata used is ancient (IC9) and maybe DOMcalibrator no longer is\nhappy with this. \n\nUpdate to newer data. DOMcalibrator is on the decline anyway.", "reporter": "blaufuss", "cc": "", "resolution": "fixed", "_ts": "1416713877111216", "component": "combo core", "summary": "a_nocompression.py test fails badly in DOMcalibrator.", "priority": "normal", "keywords": "", "time": "2011-07-20T17:28:25", "milestone": "", "owner": "blaufuss", "type": "defect" } ``` </p> </details>
1.0
a_nocompression.py test fails badly in DOMcalibrator. (Trac #294) - The WFs produced are complete craps, loads of NAN values, etc. This has been broken for a while now and gone unnoticed. The data used is ancient (IC9) and maybe DOMcalibrator no longer is happy with this. Update to newer data. DOMcalibrator is on the decline anyway. <details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/294 , reported by blaufuss and owned by blaufuss_</summary> <p> ```json { "status": "closed", "changetime": "2014-11-23T03:37:57", "description": "The WFs produced are complete craps, loads of NAN values, etc. This\nhas been broken for a while now and gone unnoticed. The\ndata used is ancient (IC9) and maybe DOMcalibrator no longer is\nhappy with this. \n\nUpdate to newer data. DOMcalibrator is on the decline anyway.", "reporter": "blaufuss", "cc": "", "resolution": "fixed", "_ts": "1416713877111216", "component": "combo core", "summary": "a_nocompression.py test fails badly in DOMcalibrator.", "priority": "normal", "keywords": "", "time": "2011-07-20T17:28:25", "milestone": "", "owner": "blaufuss", "type": "defect" } ``` </p> </details>
defect
a nocompression py test fails badly in domcalibrator trac the wfs produced are complete craps loads of nan values etc this has been broken for a while now and gone unnoticed the data used is ancient and maybe domcalibrator no longer is happy with this update to newer data domcalibrator is on the decline anyway migrated from reported by blaufuss and owned by blaufuss json status closed changetime description the wfs produced are complete craps loads of nan values etc this nhas been broken for a while now and gone unnoticed the ndata used is ancient and maybe domcalibrator no longer is nhappy with this n nupdate to newer data domcalibrator is on the decline anyway reporter blaufuss cc resolution fixed ts component combo core summary a nocompression py test fails badly in domcalibrator priority normal keywords time milestone owner blaufuss type defect
1
533,631
15,595,565,751
IssuesEvent
2021-03-18 15:01:02
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
[Coverity CID :219559] Out-of-bounds access in tests/arch/arm/arm_interrupt/src/arm_interrupt.c
Coverity Wont Fix bug priority: low
Static code scan issues found in file: https://github.com/zephyrproject-rtos/zephyr/tree/bd97359a5338b2542d19011b6d6aa1d8d1b9cc3f/tests/arch/arm/arm_interrupt/src/arm_interrupt.c Category: Memory - corruptions Function: `test_arm_interrupt` Component: Tests CID: [219559](https://scan9.coverity.com/reports.htm#v29726/p12996/mergedDefectId=219559) Details: https://github.com/zephyrproject-rtos/zephyr/blob/bd97359a5338b2542d19011b6d6aa1d8d1b9cc3f/tests/arch/arm/arm_interrupt/src/arm_interrupt.c#L307 Please fix or provide comments in coverity using the link: https://scan9.coverity.com/reports.htm#v32951/p12996. Note: This issue was created automatically. Priority was set based on classification of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
1.0
[Coverity CID :219559] Out-of-bounds access in tests/arch/arm/arm_interrupt/src/arm_interrupt.c - Static code scan issues found in file: https://github.com/zephyrproject-rtos/zephyr/tree/bd97359a5338b2542d19011b6d6aa1d8d1b9cc3f/tests/arch/arm/arm_interrupt/src/arm_interrupt.c Category: Memory - corruptions Function: `test_arm_interrupt` Component: Tests CID: [219559](https://scan9.coverity.com/reports.htm#v29726/p12996/mergedDefectId=219559) Details: https://github.com/zephyrproject-rtos/zephyr/blob/bd97359a5338b2542d19011b6d6aa1d8d1b9cc3f/tests/arch/arm/arm_interrupt/src/arm_interrupt.c#L307 Please fix or provide comments in coverity using the link: https://scan9.coverity.com/reports.htm#v32951/p12996. Note: This issue was created automatically. Priority was set based on classification of the file affected and the impact field in coverity. Assignees were set using the CODEOWNERS file.
non_defect
out of bounds access in tests arch arm arm interrupt src arm interrupt c static code scan issues found in file category memory corruptions function test arm interrupt component tests cid details please fix or provide comments in coverity using the link note this issue was created automatically priority was set based on classification of the file affected and the impact field in coverity assignees were set using the codeowners file
0
76,868
26,646,465,641
IssuesEvent
2023-01-25 10:21:07
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
Jitsi server preference not respected
T-Defect
### Steps to reproduce With the guide form https://github.com/vector-im/element-web/blob/develop/docs/jitsi.md I've tried setting things up that Element Desktop uses my own Jitsi server instead of meet.element.io. I've tried both `~/.config/Element/config.json` as well as `matrix.DOMAIN.TLD/.well-known/matrix/client`, both with no success. Depending on which source you look at sometimes its `preferred_domain`, sometimes `preferredDomain` , sometimes `jitsi` and sometimes `im.vector.riot.jitsi`. I've tried all combinations with no success. ~/.config/Element/config.json ```json { "jitsi": { "preferredDomain": "meet.DOMAIN.TLD", "preferred_domain": "meet.DOMAIN.TLD" }, "im.vector.riot.jitsi": { "preferredDomain": "meet.DOMAIN.TLD", "preferred_domain": "meet.DOMAIN.TLD" } } ``` from my nginx configuration on my homeserver ``` ... location /.well-known/matrix/client { default_type application/json; add_header Access-Control-Allow-Origin *; return 200 '{ "m.homeserver": { "base_url": "https://matrix.DOMAIN.TLD" }, "jitsi": { "preferred_domain": "meet.DOMAIN.TLD", "preferredDomain": "meet.DOMAIN.TLD"}, "im.vector.riot.jitsi": { "preferred_domain": "meet.DOMAIN.TLD", "preferredDomain": "meet.DOMAIN.TLD"}}'; } ... ``` I'm checking where the requests go to via the Dev Tools inside the Element desktop client. ### Outcome That element-desktop would use my confugred Jitsi server. ### Operating system _No response_ ### Application version Element Desktop 1.11.17 ### How did you install the app? Flathub ### Homeserver _No response_ ### Will you send logs? No
1.0
Jitsi server preference not respected - ### Steps to reproduce With the guide form https://github.com/vector-im/element-web/blob/develop/docs/jitsi.md I've tried setting things up that Element Desktop uses my own Jitsi server instead of meet.element.io. I've tried both `~/.config/Element/config.json` as well as `matrix.DOMAIN.TLD/.well-known/matrix/client`, both with no success. Depending on which source you look at sometimes its `preferred_domain`, sometimes `preferredDomain` , sometimes `jitsi` and sometimes `im.vector.riot.jitsi`. I've tried all combinations with no success. ~/.config/Element/config.json ```json { "jitsi": { "preferredDomain": "meet.DOMAIN.TLD", "preferred_domain": "meet.DOMAIN.TLD" }, "im.vector.riot.jitsi": { "preferredDomain": "meet.DOMAIN.TLD", "preferred_domain": "meet.DOMAIN.TLD" } } ``` from my nginx configuration on my homeserver ``` ... location /.well-known/matrix/client { default_type application/json; add_header Access-Control-Allow-Origin *; return 200 '{ "m.homeserver": { "base_url": "https://matrix.DOMAIN.TLD" }, "jitsi": { "preferred_domain": "meet.DOMAIN.TLD", "preferredDomain": "meet.DOMAIN.TLD"}, "im.vector.riot.jitsi": { "preferred_domain": "meet.DOMAIN.TLD", "preferredDomain": "meet.DOMAIN.TLD"}}'; } ... ``` I'm checking where the requests go to via the Dev Tools inside the Element desktop client. ### Outcome That element-desktop would use my confugred Jitsi server. ### Operating system _No response_ ### Application version Element Desktop 1.11.17 ### How did you install the app? Flathub ### Homeserver _No response_ ### Will you send logs? No
defect
jitsi server preference not respected steps to reproduce with the guide form i ve tried setting things up that element desktop uses my own jitsi server instead of meet element io i ve tried both config element config json as well as matrix domain tld well known matrix client both with no success depending on which source you look at sometimes its preferred domain sometimes preferreddomain sometimes jitsi and sometimes im vector riot jitsi i ve tried all combinations with no success config element config json json jitsi preferreddomain meet domain tld preferred domain meet domain tld im vector riot jitsi preferreddomain meet domain tld preferred domain meet domain tld from my nginx configuration on my homeserver location well known matrix client default type application json add header access control allow origin return m homeserver base url jitsi preferred domain meet domain tld preferreddomain meet domain tld im vector riot jitsi preferred domain meet domain tld preferreddomain meet domain tld i m checking where the requests go to via the dev tools inside the element desktop client outcome that element desktop would use my confugred jitsi server operating system no response application version element desktop how did you install the app flathub homeserver no response will you send logs no
1
618,925
19,491,760,882
IssuesEvent
2021-12-27 07:57:18
aave/aave-ui
https://api.github.com/repos/aave/aave-ui
reopened
V3 Dashboard
priority:high V3 txflow
<!-- Provide a general summary of the feature in the Title above --> From yesterdays meeting we need to polish some of the new dashboard UI. - Remove deposit and borrow white boxes when there are no positions. - Update text from "Assets to deposit" to "Available in your wallet" - Assets on borrow side should always be shown, regardless if you have a position or not. - Arbitrum banner should be able to be closed via "X". This should also be moved if possible - add swap instead of deposit on the list of supply positions - When an asset is partially deposited and partially in the wallet, we show it in both top and bottom part; the both part will only have withdraw/swap, the bottom part will only have deposit/details <!-- Thank you very much for contributing to AAVE ui by creating an issue! ❤️ To avoid duplicate issues we ask you to check off the following list. --> <!-- Checked checkbox should look like this: [x] --> - [x] I have searched the [issues](https://github.com/aave/aave-ui/issues) of this repository and believe that this is not a duplicate. ## Summary 💡 <!-- Describe how it should work. --> ## Examples 🌈 <img width="1423" alt="Screen Shot 2021-12-24 at 7 27 45 AM" src="https://user-images.githubusercontent.com/8342048/147353305-a7766ee8-4fb1-47c7-98e6-3694adf99a64.png"> - Note: Layout seemed broken with isolation mode and borrow positions <img width="1436" alt="Screen Shot 2021-12-24 at 7 35 45 AM" src="https://user-images.githubusercontent.com/8342048/147353420-ecae2b8a-004c-4603-adc4-7c7537c5647e.png"> <!-- Provide examples of other platforms solutions if feasible. --> ## Motivation 🔦 <!-- What are you trying to accomplish? How has the lack of this feature affected you? If there's a related [governance discussion](https://governance.aave.com) please link it! -->
1.0
V3 Dashboard - <!-- Provide a general summary of the feature in the Title above --> From yesterdays meeting we need to polish some of the new dashboard UI. - Remove deposit and borrow white boxes when there are no positions. - Update text from "Assets to deposit" to "Available in your wallet" - Assets on borrow side should always be shown, regardless if you have a position or not. - Arbitrum banner should be able to be closed via "X". This should also be moved if possible - add swap instead of deposit on the list of supply positions - When an asset is partially deposited and partially in the wallet, we show it in both top and bottom part; the both part will only have withdraw/swap, the bottom part will only have deposit/details <!-- Thank you very much for contributing to AAVE ui by creating an issue! ❤️ To avoid duplicate issues we ask you to check off the following list. --> <!-- Checked checkbox should look like this: [x] --> - [x] I have searched the [issues](https://github.com/aave/aave-ui/issues) of this repository and believe that this is not a duplicate. ## Summary 💡 <!-- Describe how it should work. --> ## Examples 🌈 <img width="1423" alt="Screen Shot 2021-12-24 at 7 27 45 AM" src="https://user-images.githubusercontent.com/8342048/147353305-a7766ee8-4fb1-47c7-98e6-3694adf99a64.png"> - Note: Layout seemed broken with isolation mode and borrow positions <img width="1436" alt="Screen Shot 2021-12-24 at 7 35 45 AM" src="https://user-images.githubusercontent.com/8342048/147353420-ecae2b8a-004c-4603-adc4-7c7537c5647e.png"> <!-- Provide examples of other platforms solutions if feasible. --> ## Motivation 🔦 <!-- What are you trying to accomplish? How has the lack of this feature affected you? If there's a related [governance discussion](https://governance.aave.com) please link it! -->
non_defect
dashboard from yesterdays meeting we need to polish some of the new dashboard ui remove deposit and borrow white boxes when there are no positions update text from assets to deposit to available in your wallet assets on borrow side should always be shown regardless if you have a position or not arbitrum banner should be able to be closed via x this should also be moved if possible add swap instead of deposit on the list of supply positions when an asset is partially deposited and partially in the wallet we show it in both top and bottom part the both part will only have withdraw swap the bottom part will only have deposit details thank you very much for contributing to aave ui by creating an issue ❤️ to avoid duplicate issues we ask you to check off the following list i have searched the of this repository and believe that this is not a duplicate summary 💡 examples 🌈 img width alt screen shot at am src note layout seemed broken with isolation mode and borrow positions img width alt screen shot at am src provide examples of other platforms solutions if feasible motivation 🔦 what are you trying to accomplish how has the lack of this feature affected you if there s a related please link it
0
62,913
17,261,471,347
IssuesEvent
2021-07-22 08:16:07
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Couldn't terminate video call
A-Jitsi A-VoIP P1 S-Major T-Defect
What happened: - Alice tried to use the 'easter egg' screenshare feature - didn't seem to work - My Element sessions (two of them) rang and the event said it as a voice call - My Element sessions rang out - Alice set up a jitsi widget - I initiated a 1:1 video call very shortly after - Call connected successfully, great success - I clicked 'End conference' - Call would not end - Alice couldn't end the call either - Awkward permanent video connection
1.0
Couldn't terminate video call - What happened: - Alice tried to use the 'easter egg' screenshare feature - didn't seem to work - My Element sessions (two of them) rang and the event said it as a voice call - My Element sessions rang out - Alice set up a jitsi widget - I initiated a 1:1 video call very shortly after - Call connected successfully, great success - I clicked 'End conference' - Call would not end - Alice couldn't end the call either - Awkward permanent video connection
defect
couldn t terminate video call what happened alice tried to use the easter egg screenshare feature didn t seem to work my element sessions two of them rang and the event said it as a voice call my element sessions rang out alice set up a jitsi widget i initiated a video call very shortly after call connected successfully great success i clicked end conference call would not end alice couldn t end the call either awkward permanent video connection
1
61,315
17,023,665,026
IssuesEvent
2021-07-03 03:10:58
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Gosmore on Windows CE/Mobile: Shows too few search results
Component: gosmore Priority: major Resolution: wontfix Type: defect
**[Submitted to the original trac issue database at 5.53pm, Saturday, 25th December 2010]** Non trivial routing with gosmore on wince (tested on WinCE 5.0) is almost impossible because scrolling is not implemented. The natural way to set an end point is to use the search box. On a small screen, typically only 3 entries are shown, the lower half of the screen being covered by the virtual keyboard. There seems to be no way to scroll down to find more distant places. Trying to zoom out on the main map to manually centre on a distant end point is pretty hopeless, since gosmore becomes impossibly slow at low zoom ratios. This coupled with the fact that gosmore then typically loses the end point when the zoom level is changed (etc) mentioned in another ticket means that it cannot be used for more than very local routing on portable devices. At least while they are running wince.
1.0
Gosmore on Windows CE/Mobile: Shows too few search results - **[Submitted to the original trac issue database at 5.53pm, Saturday, 25th December 2010]** Non trivial routing with gosmore on wince (tested on WinCE 5.0) is almost impossible because scrolling is not implemented. The natural way to set an end point is to use the search box. On a small screen, typically only 3 entries are shown, the lower half of the screen being covered by the virtual keyboard. There seems to be no way to scroll down to find more distant places. Trying to zoom out on the main map to manually centre on a distant end point is pretty hopeless, since gosmore becomes impossibly slow at low zoom ratios. This coupled with the fact that gosmore then typically loses the end point when the zoom level is changed (etc) mentioned in another ticket means that it cannot be used for more than very local routing on portable devices. At least while they are running wince.
defect
gosmore on windows ce mobile shows too few search results non trivial routing with gosmore on wince tested on wince is almost impossible because scrolling is not implemented the natural way to set an end point is to use the search box on a small screen typically only entries are shown the lower half of the screen being covered by the virtual keyboard there seems to be no way to scroll down to find more distant places trying to zoom out on the main map to manually centre on a distant end point is pretty hopeless since gosmore becomes impossibly slow at low zoom ratios this coupled with the fact that gosmore then typically loses the end point when the zoom level is changed etc mentioned in another ticket means that it cannot be used for more than very local routing on portable devices at least while they are running wince
1
40,907
10,220,881,027
IssuesEvent
2019-08-15 22:55:46
idaholab/moose
https://api.github.com/repos/idaholab/moose
opened
Compiler warnings about MOOSE dummies
C: MOOSE P: minor T: defect
## Bug Description Every Registry.h macro invocation generates an 11 line long compiler warning for me, e.g.: ``` /home/roystgnr/git/moose-fresh/framework/build/header_symlinks/Registry.h:47:28: warning: ‘dummyvar_for_registering_obj_SmoothMeshGenerator19’ defined but not used [-Wunused-variable] static char combineNames(dummyvar_for_registering_obj_##classname, __LINE__) = \ ^ /home/roystgnr/git/moose-fresh/framework/build/header_symlinks/Registry.h:17:29: note: in definition of macro ‘combineNames1’ #define combineNames1(X, Y) X##Y ^ /home/roystgnr/git/moose-fresh/framework/build/header_symlinks/Registry.h:47:15: note: in expansion of macro ‘combineNames’ static char combineNames(dummyvar_for_registering_obj_##classname, __LINE__) = \ ^ /home/roystgnr/git/moose-fresh/framework/src/meshgenerators/SmoothMeshGenerator.C:19:1: note: in expansion of macro ‘registerMooseObject’ registerMooseObject("MooseApp", SmoothMeshGenerator); ``` ## Steps to Reproduce Build MOOSE with the -Wunused-variable compiler flag or with one of the many flags that incorporates it. ## Impact This is just an annoyance, but it's a voluminous one: tens of thousands of lines of warning text in a typical MOOSE build. What's the point of the static char here?
1.0
Compiler warnings about MOOSE dummies - ## Bug Description Every Registry.h macro invocation generates an 11 line long compiler warning for me, e.g.: ``` /home/roystgnr/git/moose-fresh/framework/build/header_symlinks/Registry.h:47:28: warning: ‘dummyvar_for_registering_obj_SmoothMeshGenerator19’ defined but not used [-Wunused-variable] static char combineNames(dummyvar_for_registering_obj_##classname, __LINE__) = \ ^ /home/roystgnr/git/moose-fresh/framework/build/header_symlinks/Registry.h:17:29: note: in definition of macro ‘combineNames1’ #define combineNames1(X, Y) X##Y ^ /home/roystgnr/git/moose-fresh/framework/build/header_symlinks/Registry.h:47:15: note: in expansion of macro ‘combineNames’ static char combineNames(dummyvar_for_registering_obj_##classname, __LINE__) = \ ^ /home/roystgnr/git/moose-fresh/framework/src/meshgenerators/SmoothMeshGenerator.C:19:1: note: in expansion of macro ‘registerMooseObject’ registerMooseObject("MooseApp", SmoothMeshGenerator); ``` ## Steps to Reproduce Build MOOSE with the -Wunused-variable compiler flag or with one of the many flags that incorporates it. ## Impact This is just an annoyance, but it's a voluminous one: tens of thousands of lines of warning text in a typical MOOSE build. What's the point of the static char here?
defect
compiler warnings about moose dummies bug description every registry h macro invocation generates an line long compiler warning for me e g home roystgnr git moose fresh framework build header symlinks registry h warning ‘dummyvar for registering obj ’ defined but not used static char combinenames dummyvar for registering obj classname line home roystgnr git moose fresh framework build header symlinks registry h note in definition of macro ‘ ’ define x y x y home roystgnr git moose fresh framework build header symlinks registry h note in expansion of macro ‘combinenames’ static char combinenames dummyvar for registering obj classname line home roystgnr git moose fresh framework src meshgenerators smoothmeshgenerator c note in expansion of macro ‘registermooseobject’ registermooseobject mooseapp smoothmeshgenerator steps to reproduce build moose with the wunused variable compiler flag or with one of the many flags that incorporates it impact this is just an annoyance but it s a voluminous one tens of thousands of lines of warning text in a typical moose build what s the point of the static char here
1
4,526
2,610,112,555
IssuesEvent
2015-02-26 18:34:53
chrsmith/scribefire-chrome
https://api.github.com/repos/chrsmith/scribefire-chrome
closed
Scribefire Download link reports "file deprecated"
auto-migrated Priority-Medium Type-Defect
``` What's the problem? Scribefire Download Link on code.google.com reports, "This file is deprecated. Another file would probably be better." I don't know what this means, but I trust Google. When it offers that I should download "another file" I would do so but do not know from where to find it since no other alternative link is presented. What version of ScribeFire for Chrome are you running? None, yet. ``` ----- Original issue reported on code.google.com by `amoswhite3` on 19 May 2010 at 5:52
1.0
Scribefire Download link reports "file deprecated" - ``` What's the problem? Scribefire Download Link on code.google.com reports, "This file is deprecated. Another file would probably be better." I don't know what this means, but I trust Google. When it offers that I should download "another file" I would do so but do not know from where to find it since no other alternative link is presented. What version of ScribeFire for Chrome are you running? None, yet. ``` ----- Original issue reported on code.google.com by `amoswhite3` on 19 May 2010 at 5:52
defect
scribefire download link reports file deprecated what s the problem scribefire download link on code google com reports this file is deprecated another file would probably be better i don t know what this means but i trust google when it offers that i should download another file i would do so but do not know from where to find it since no other alternative link is presented what version of scribefire for chrome are you running none yet original issue reported on code google com by on may at
1
349,811
10,473,900,097
IssuesEvent
2019-09-23 13:33:03
craftercms/craftercms
https://api.github.com/repos/craftercms/craftercms
opened
[engine] Provide cache warming and cache double buffer.
enhancement priority: high
Engine should be able to do cache warm-up after a cache clear has been requested (configurable) and be able to do this warm-up in the background without actually clearing the current cache (double buffer of caches). This is specially important for serverless.
1.0
[engine] Provide cache warming and cache double buffer. - Engine should be able to do cache warm-up after a cache clear has been requested (configurable) and be able to do this warm-up in the background without actually clearing the current cache (double buffer of caches). This is specially important for serverless.
non_defect
provide cache warming and cache double buffer engine should be able to do cache warm up after a cache clear has been requested configurable and be able to do this warm up in the background without actually clearing the current cache double buffer of caches this is specially important for serverless
0
650,921
21,443,099,153
IssuesEvent
2022-04-25 01:09:31
kubernetes-sigs/oci-proxy
https://api.github.com/repos/kubernetes-sigs/oci-proxy
closed
Add `preset-use-new-registry` to various jobs
priority/important-longterm sig/k8s-infra
In order to add some production load testing on the registry.k8s.io redirection, a request from @dims is to add the label `preset-use-new-registry=true` so that some e2e tests that pull images can pull through the new registry ref: https://kubernetes.slack.com/archives/CCK68P2Q2/p1647896105383489
1.0
Add `preset-use-new-registry` to various jobs - In order to add some production load testing on the registry.k8s.io redirection, a request from @dims is to add the label `preset-use-new-registry=true` so that some e2e tests that pull images can pull through the new registry ref: https://kubernetes.slack.com/archives/CCK68P2Q2/p1647896105383489
non_defect
add preset use new registry to various jobs in order to add some production load testing on the registry io redirection a request from dims is to add the label preset use new registry true so that some tests that pull images can pull through the new registry ref
0
49,075
13,185,216,975
IssuesEvent
2020-08-12 20:57:29
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
I3File does not correctly follow Python iterator interface (Trac #688)
Incomplete Migration Migrated from Trac dataio defect
<details> <summary><em>Migrated from https://code.icecube.wisc.edu/ticket/688 , reported by sjackso and owned by sjackso</em></summary> <p> ```json { "status": "closed", "changetime": "2014-03-22T04:28:38", "description": "The Python binding for an I3File, implemented in C++ as I3SequentialFile, provides methods `next()` and `__iter__()` to implement the Python iterator interface. However, it does not support the iterator interface correctly, because an I3File is '''both''' a container and an iterator. A correct implementation would provide a separate iterator class that implemented `next()` and `__iter__()`, while I3File itself would only provide `__iter__()`.\n\nReference: http://docs.python.org/library/stdtypes.html#iterator-types\n\nIn rare cases, this issue can cause unexpected iterator behavior. For example:\n\n{{{\n1 it = iter(i3file)\n2 frame1 = it.next()\n3 for frame in it:\n4 # attempt to act on second and all subsequent frames...\n}}}\n\nOn the first run through the loop, at line 4, `frame` will equal `frame1`.\n\nI'm giving this a low priority, but figured it should be documented.", "reporter": "sjackso", "cc": "", "resolution": "fixed", "_ts": "1395462518000000", "component": "dataio", "summary": "I3File does not correctly follow Python iterator interface", "priority": "minor", "keywords": "I3File I3SequentialFile iterator", "time": "2012-09-27T19:52:15", "milestone": "", "owner": "sjackso", "type": "defect" } ``` </p> </details>
1.0
I3File does not correctly follow Python iterator interface (Trac #688) - <details> <summary><em>Migrated from https://code.icecube.wisc.edu/ticket/688 , reported by sjackso and owned by sjackso</em></summary> <p> ```json { "status": "closed", "changetime": "2014-03-22T04:28:38", "description": "The Python binding for an I3File, implemented in C++ as I3SequentialFile, provides methods `next()` and `__iter__()` to implement the Python iterator interface. However, it does not support the iterator interface correctly, because an I3File is '''both''' a container and an iterator. A correct implementation would provide a separate iterator class that implemented `next()` and `__iter__()`, while I3File itself would only provide `__iter__()`.\n\nReference: http://docs.python.org/library/stdtypes.html#iterator-types\n\nIn rare cases, this issue can cause unexpected iterator behavior. For example:\n\n{{{\n1 it = iter(i3file)\n2 frame1 = it.next()\n3 for frame in it:\n4 # attempt to act on second and all subsequent frames...\n}}}\n\nOn the first run through the loop, at line 4, `frame` will equal `frame1`.\n\nI'm giving this a low priority, but figured it should be documented.", "reporter": "sjackso", "cc": "", "resolution": "fixed", "_ts": "1395462518000000", "component": "dataio", "summary": "I3File does not correctly follow Python iterator interface", "priority": "minor", "keywords": "I3File I3SequentialFile iterator", "time": "2012-09-27T19:52:15", "milestone": "", "owner": "sjackso", "type": "defect" } ``` </p> </details>
defect
does not correctly follow python iterator interface trac migrated from reported by sjackso and owned by sjackso json status closed changetime description the python binding for an implemented in c as provides methods next and iter to implement the python iterator interface however it does not support the iterator interface correctly because an is both a container and an iterator a correct implementation would provide a separate iterator class that implemented next and iter while itself would only provide iter n nreference rare cases this issue can cause unexpected iterator behavior for example n n it iter it next for frame in it attempt to act on second and all subsequent frames n n non the first run through the loop at line frame will equal n ni m giving this a low priority but figured it should be documented reporter sjackso cc resolution fixed ts component dataio summary does not correctly follow python iterator interface priority minor keywords iterator time milestone owner sjackso type defect
1
33,613
7,180,497,375
IssuesEvent
2018-01-31 23:34:19
idaholab/moose
https://api.github.com/repos/idaholab/moose
opened
Close Vectors / Matrices On Exception
C: Modules P: normal T: defect
## Rationale When an exception occurs it could leave a vector / matrix in an invalid state. We should close them. Not doing so is currently causing a few tests to fail. ## Description In the error handler (`catch`) we should go through all vectors / matrices and close them when an exception is thrown. ## Impact Objects can be in wrong state after throwing the exception currently.
1.0
Close Vectors / Matrices On Exception - ## Rationale When an exception occurs it could leave a vector / matrix in an invalid state. We should close them. Not doing so is currently causing a few tests to fail. ## Description In the error handler (`catch`) we should go through all vectors / matrices and close them when an exception is thrown. ## Impact Objects can be in wrong state after throwing the exception currently.
defect
close vectors matrices on exception rationale when an exception occurs it could leave a vector matrix in an invalid state we should close them not doing so is currently causing a few tests to fail description in the error handler catch we should go through all vectors matrices and close them when an exception is thrown impact objects can be in wrong state after throwing the exception currently
1
364,443
25,491,240,488
IssuesEvent
2022-11-27 04:24:23
octopusdream/infra
https://api.github.com/repos/octopusdream/infra
closed
Thanos Research
documentation
Prometheus HA 를 위한 Thanos 리서치 우리 시스템에 적용할 수 있는지, 효용이 있는지 조사 - [ ] 리서치 - [ ] 문서작성
1.0
Thanos Research - Prometheus HA 를 위한 Thanos 리서치 우리 시스템에 적용할 수 있는지, 효용이 있는지 조사 - [ ] 리서치 - [ ] 문서작성
non_defect
thanos research prometheus ha 를 위한 thanos 리서치 우리 시스템에 적용할 수 있는지 효용이 있는지 조사 리서치 문서작성
0
7,930
2,611,067,408
IssuesEvent
2015-02-27 00:31:29
alistairreilly/andors-trail
https://api.github.com/repos/alistairreilly/andors-trail
closed
RoLS - regeneration
auto-migrated Milestone-0.6.10 Type-Defect
``` Equip the RoLS, wait 5 seconds, then enter a new map. As you enter the new map the Role's regeneration will fire and will cause the entire map to go black except for the tile your character is on. What steps will reproduce the problem? 1. 2. 3. What is the expected output? What do you see instead? What version of the product are you using? On what device? 0.6.9 DROID x 2.3 Please provide any additional information below. ``` Original issue reported on code.google.com by `jgemmajr@gmail.com` on 10 Aug 2011 at 2:10
1.0
RoLS - regeneration - ``` Equip the RoLS, wait 5 seconds, then enter a new map. As you enter the new map the Role's regeneration will fire and will cause the entire map to go black except for the tile your character is on. What steps will reproduce the problem? 1. 2. 3. What is the expected output? What do you see instead? What version of the product are you using? On what device? 0.6.9 DROID x 2.3 Please provide any additional information below. ``` Original issue reported on code.google.com by `jgemmajr@gmail.com` on 10 Aug 2011 at 2:10
defect
rols regeneration equip the rols wait seconds then enter a new map as you enter the new map the role s regeneration will fire and will cause the entire map to go black except for the tile your character is on what steps will reproduce the problem what is the expected output what do you see instead what version of the product are you using on what device droid x please provide any additional information below original issue reported on code google com by jgemmajr gmail com on aug at
1
320,750
27,455,564,809
IssuesEvent
2023-03-02 21:02:40
w3c/csswg-drafts
https://api.github.com/repos/w3c/csswg-drafts
closed
[css-color-4] `hsl()` modern syntax allows `<number>` but seems to be lacking details and WPT tests
css-color-4 Needs Edits Needs Testcase (WPT)
ERROR: type should be string, got "https://drafts.csswg.org/css-color-4/#the-hsl-notation\r\n\r\n```\r\n<modern-hsl-syntax> = hsl( \r\n [<hue> | none] \r\n [<percentage> | <number> | none] \r\n [<percentage> | <number> | none] \r\n [ / [<alpha-value> | none] ]? )\r\n```\r\n\r\nThe prose and the examples don't give any indication or even mention `<number>`.\r\nI also couldn't find any WPT tests, but there are many and I might be overlooking cases where `%` was omitted.\r\n\r\nI am assuming this was a recent addition because no browser implements support for numbers in `s` and/or `l`\r\n\r\n```css\r\n* {\r\n background-color: hsl(50deg 0.5 0.5);\r\n background-color: hsl(50deg 50 50);\r\n}\r\n```\r\n\r\nAll other color functions that support numbers and percentages have the percentage range in prose or in a section :\r\n\r\n```\r\nPercent reference range \tfor L: 0% = 0.0, 100% = 1.0\r\n for C: 0% = 0.0 100% = 0.4\r\n```"
1.0
[css-color-4] `hsl()` modern syntax allows `<number>` but seems to be lacking details and WPT tests - https://drafts.csswg.org/css-color-4/#the-hsl-notation ``` <modern-hsl-syntax> = hsl( [<hue> | none] [<percentage> | <number> | none] [<percentage> | <number> | none] [ / [<alpha-value> | none] ]? ) ``` The prose and the examples don't give any indication or even mention `<number>`. I also couldn't find any WPT tests, but there are many and I might be overlooking cases where `%` was omitted. I am assuming this was a recent addition because no browser implements support for numbers in `s` and/or `l` ```css * { background-color: hsl(50deg 0.5 0.5); background-color: hsl(50deg 50 50); } ``` All other color functions that support numbers and percentages have the percentage range in prose or in a section : ``` Percent reference range for L: 0% = 0.0, 100% = 1.0 for C: 0% = 0.0 100% = 0.4 ```
non_defect
hsl modern syntax allows but seems to be lacking details and wpt tests hsl the prose and the examples don t give any indication or even mention i also couldn t find any wpt tests but there are many and i might be overlooking cases where was omitted i am assuming this was a recent addition because no browser implements support for numbers in s and or l css background color hsl background color hsl all other color functions that support numbers and percentages have the percentage range in prose or in a section percent reference range for l for c
0
9,118
2,615,132,358
IssuesEvent
2015-03-01 06:02:34
chrsmith/google-api-java-client
https://api.github.com/repos/chrsmith/google-api-java-client
closed
NoSuchMethodError exception for com.google.api.client.auth.oauth2.AuthorizationCodeRequestUrl.setScopes(Ljava/util/Collection;)
auto-migrated Priority-Medium Type-Defect
``` Version of google-api-java-client (e.g. 1.15.0-rc)? 1.15.0-rc Java environment (e.g. Java 6, Android 2.3, App Engine)? Java 6 and Java 7 Describe the problem. java.lang.NoSuchMethodError: com.google.api.client.auth.oauth2.AuthorizationCodeRequestUrl.setScopes(Ljava/ut il/Collection;)Lcom/google/api/client/auth/oauth2/AuthorizationCodeRequestUrl; The problematic line of code: AuthorizationCodeRequestUrl authorizationCodeURL=new AuthorizationCodeRequestUrl(URL, ID); authorizationCodeURL.setScopes(scopes); //Scopes is an array list of strings Thing is, this worked in a previous version of api using GoogleAuthorizationCodeRequestUrl class, which is deprecated now. How would you expect it to be fixed? Not sure, the method in question (setScopes) is there, the correct parameters for it are there. ``` Original issue reported on code.google.com by `DarkR4z...@gmail.com` on 14 May 2013 at 2:17
1.0
NoSuchMethodError exception for com.google.api.client.auth.oauth2.AuthorizationCodeRequestUrl.setScopes(Ljava/util/Collection;) - ``` Version of google-api-java-client (e.g. 1.15.0-rc)? 1.15.0-rc Java environment (e.g. Java 6, Android 2.3, App Engine)? Java 6 and Java 7 Describe the problem. java.lang.NoSuchMethodError: com.google.api.client.auth.oauth2.AuthorizationCodeRequestUrl.setScopes(Ljava/ut il/Collection;)Lcom/google/api/client/auth/oauth2/AuthorizationCodeRequestUrl; The problematic line of code: AuthorizationCodeRequestUrl authorizationCodeURL=new AuthorizationCodeRequestUrl(URL, ID); authorizationCodeURL.setScopes(scopes); //Scopes is an array list of strings Thing is, this worked in a previous version of api using GoogleAuthorizationCodeRequestUrl class, which is deprecated now. How would you expect it to be fixed? Not sure, the method in question (setScopes) is there, the correct parameters for it are there. ``` Original issue reported on code.google.com by `DarkR4z...@gmail.com` on 14 May 2013 at 2:17
defect
nosuchmethoderror exception for com google api client auth authorizationcoderequesturl setscopes ljava util collection version of google api java client e g rc rc java environment e g java android app engine java and java describe the problem java lang nosuchmethoderror com google api client auth authorizationcoderequesturl setscopes ljava ut il collection lcom google api client auth authorizationcoderequesturl the problematic line of code authorizationcoderequesturl authorizationcodeurl new authorizationcoderequesturl url id authorizationcodeurl setscopes scopes scopes is an array list of strings thing is this worked in a previous version of api using googleauthorizationcoderequesturl class which is deprecated now how would you expect it to be fixed not sure the method in question setscopes is there the correct parameters for it are there original issue reported on code google com by gmail com on may at
1
38,027
8,637,320,046
IssuesEvent
2018-11-23 10:50:55
cakephp/cakephp
https://api.github.com/repos/cakephp/cakephp
closed
NullEngine does not respect abstract method's return type declaration
Defect
This is a (multiple allowed): * [x] bug * [ ] enhancement * [ ] feature-discussion (RFC) * CakePHP Version: 3.6.13 * Platform and Target: PHP's build-in webserver. ### What you did Trying to write to cache with caching disabled. ```php <?php use Cake\Cache\Cache; Cache::disable(); $success = Cache::write('foo', 'bar'); ``` ### What happened `$success` value is **null**. ### What you expected to happen `$success` value should be **boolean false**.
1.0
NullEngine does not respect abstract method's return type declaration - This is a (multiple allowed): * [x] bug * [ ] enhancement * [ ] feature-discussion (RFC) * CakePHP Version: 3.6.13 * Platform and Target: PHP's build-in webserver. ### What you did Trying to write to cache with caching disabled. ```php <?php use Cake\Cache\Cache; Cache::disable(); $success = Cache::write('foo', 'bar'); ``` ### What happened `$success` value is **null**. ### What you expected to happen `$success` value should be **boolean false**.
defect
nullengine does not respect abstract method s return type declaration this is a multiple allowed bug enhancement feature discussion rfc cakephp version platform and target php s build in webserver what you did trying to write to cache with caching disabled php php use cake cache cache cache disable success cache write foo bar what happened success value is null what you expected to happen success value should be boolean false
1
3,077
2,607,983,017
IssuesEvent
2015-02-26 00:50:25
chrsmithdemos/zen-coding
https://api.github.com/repos/chrsmithdemos/zen-coding
closed
Zen HTML bundle conflicts with default Textmate HTML bundle
auto-migrated Priority-Medium Type-Defect
``` The Zen HTML bundle duplicates many shortcuts already used by Textmate's default HTML bundle. This is silly, especially considering this is a bundle created exclusively for Textmate. Disabling the default HTML bundle is not an acceptable solution. It contains loads of key functions not contained in the Zen bundle. Duplicate shortcuts should be removed from the Zen HTML Bundle. Or the Zen bundle should integrate all the functionality of the default bundle. Otherwise it's just a mess. ``` ----- Original issue reported on code.google.com by `Featherodd@gmail.com` on 20 Mar 2011 at 8:16 * Merged into: #79
1.0
Zen HTML bundle conflicts with default Textmate HTML bundle - ``` The Zen HTML bundle duplicates many shortcuts already used by Textmate's default HTML bundle. This is silly, especially considering this is a bundle created exclusively for Textmate. Disabling the default HTML bundle is not an acceptable solution. It contains loads of key functions not contained in the Zen bundle. Duplicate shortcuts should be removed from the Zen HTML Bundle. Or the Zen bundle should integrate all the functionality of the default bundle. Otherwise it's just a mess. ``` ----- Original issue reported on code.google.com by `Featherodd@gmail.com` on 20 Mar 2011 at 8:16 * Merged into: #79
defect
zen html bundle conflicts with default textmate html bundle the zen html bundle duplicates many shortcuts already used by textmate s default html bundle this is silly especially considering this is a bundle created exclusively for textmate disabling the default html bundle is not an acceptable solution it contains loads of key functions not contained in the zen bundle duplicate shortcuts should be removed from the zen html bundle or the zen bundle should integrate all the functionality of the default bundle otherwise it s just a mess original issue reported on code google com by featherodd gmail com on mar at merged into
1
194,430
14,678,124,682
IssuesEvent
2020-12-31 02:06:37
github-vet/rangeloop-pointer-findings
https://api.github.com/repos/github-vet/rangeloop-pointer-findings
closed
dylhunn/dragontooth: transtable/transtable_test.go; 10 LoC
fresh test tiny
Found a possible issue in [dylhunn/dragontooth](https://www.github.com/dylhunn/dragontooth) at [transtable/transtable_test.go](https://github.com/dylhunn/dragontooth/blob/011dee96b5442d62f7af9b9d9dd0c3a3083f886d/transtable/transtable_test.go#L32-L41) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > [Click here to see the code in its original context.](https://github.com/dylhunn/dragontooth/blob/011dee96b5442d62f7af9b9d9dd0c3a3083f886d/transtable/transtable_test.go#L32-L41) <details> <summary>Click here to show the 10 line(s) of Go which triggered the analyzer.</summary> ```go for k, mv := range movesMap { b := dragontoothmg.ParseFen(k) Put(&b, mv, -30, -6, LowerBound) found, resmove, reseval, resdepth, restype := Get(&b) if !found || resmove != mv || reseval != -30 || resdepth != -6 || restype != LowerBound { t.Error("Simple ttable test failed. \nPut data:", b.ToFen(), &mv, -30, 6, Exact, "\n", "Fetched data:", found, &resmove, reseval, resdepth, restype) } } ``` </details> <details> <summary>Click here to show extra information the analyzer produced.</summary> ``` No path was found through the callgraph that could lead to a function which writes a pointer argument. No path was found through the callgraph that could lead to a function which passes a pointer to third-party code. root signature {Error 13} was not found in the callgraph; reference was passed directly to third-party code ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 011dee96b5442d62f7af9b9d9dd0c3a3083f886d
1.0
dylhunn/dragontooth: transtable/transtable_test.go; 10 LoC - Found a possible issue in [dylhunn/dragontooth](https://www.github.com/dylhunn/dragontooth) at [transtable/transtable_test.go](https://github.com/dylhunn/dragontooth/blob/011dee96b5442d62f7af9b9d9dd0c3a3083f886d/transtable/transtable_test.go#L32-L41) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > [Click here to see the code in its original context.](https://github.com/dylhunn/dragontooth/blob/011dee96b5442d62f7af9b9d9dd0c3a3083f886d/transtable/transtable_test.go#L32-L41) <details> <summary>Click here to show the 10 line(s) of Go which triggered the analyzer.</summary> ```go for k, mv := range movesMap { b := dragontoothmg.ParseFen(k) Put(&b, mv, -30, -6, LowerBound) found, resmove, reseval, resdepth, restype := Get(&b) if !found || resmove != mv || reseval != -30 || resdepth != -6 || restype != LowerBound { t.Error("Simple ttable test failed. \nPut data:", b.ToFen(), &mv, -30, 6, Exact, "\n", "Fetched data:", found, &resmove, reseval, resdepth, restype) } } ``` </details> <details> <summary>Click here to show extra information the analyzer produced.</summary> ``` No path was found through the callgraph that could lead to a function which writes a pointer argument. No path was found through the callgraph that could lead to a function which passes a pointer to third-party code. root signature {Error 13} was not found in the callgraph; reference was passed directly to third-party code ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 011dee96b5442d62f7af9b9d9dd0c3a3083f886d
non_defect
dylhunn dragontooth transtable transtable test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message click here to show the line s of go which triggered the analyzer go for k mv range movesmap b dragontoothmg parsefen k put b mv lowerbound found resmove reseval resdepth restype get b if found resmove mv reseval resdepth restype lowerbound t error simple ttable test failed nput data b tofen mv exact n fetched data found resmove reseval resdepth restype click here to show extra information the analyzer produced no path was found through the callgraph that could lead to a function which writes a pointer argument no path was found through the callgraph that could lead to a function which passes a pointer to third party code root signature error was not found in the callgraph reference was passed directly to third party code leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
0
71,959
23,869,006,827
IssuesEvent
2022-09-07 13:28:55
jOOQ/jOOQ
https://api.github.com/repos/jOOQ/jOOQ
closed
Cannot fetch inline NULL value as array type in HSQLDB
T: Defect C: Functionality C: DB: HSQLDB P: Medium E: All Editions
It seems we can't do this with HSQLDB: ```java ctx.select(inline(null, String[].class)).fetch(); ``` We're getting: ``` Exception in thread "main" org.jooq.exception.DataAccessException: SQL [select null from (values(1)) as dual(dual) -- SQL rendered with a free trial version of jOOQ 3.15.0-SNAPSHOT]; Error while reading field: null, at JDBC index: 1 at org.jooq_3.15.0-SNAPSHOT.HSQLDB.debug(Unknown Source) at org.jooq_3.15.0-SNAPSHOT.HSQLDB.debug(Unknown Source) at org.jooq.impl.Tools.translate(Tools.java:2877) at org.jooq.impl.DefaultExecuteContext.sqlException(DefaultExecuteContext.java:695) at org.jooq.impl.CursorImpl$CursorIterator.fetchNext(CursorImpl.java:1436) at org.jooq.impl.CursorImpl$CursorIterator.hasNext(CursorImpl.java:1397) at org.jooq.impl.CursorImpl.fetchNext(CursorImpl.java:217) at org.jooq.impl.AbstractCursor.fetch(AbstractCursor.java:182) at org.jooq.impl.AbstractCursor.fetch(AbstractCursor.java:93) at org.jooq.impl.AbstractResultQuery.execute(AbstractResultQuery.java:289) at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:374) at org.jooq.impl.AbstractResultQuery.fetch(AbstractResultQuery.java:314) at org.jooq.impl.SelectImpl.fetch(SelectImpl.java:2818) at org.jooq.testscripts.JDBC.main(JDBC.java:25) Caused by: java.sql.SQLException: Error while reading field: null, at JDBC index: 1 at org.jooq.impl.CursorImpl$CursorIterator$CursorRecordInitialiser.setValue(CursorImpl.java:1548) at org.jooq.impl.CursorImpl$CursorIterator$CursorRecordInitialiser.apply(CursorImpl.java:1497) at org.jooq.impl.CursorImpl$CursorIterator$CursorRecordInitialiser.apply(CursorImpl.java:1) at org.jooq.impl.RecordDelegate.operate(RecordDelegate.java:145) at org.jooq.impl.CursorImpl$CursorIterator.fetchNext(CursorImpl.java:1421) ... 9 more Caused by: java.sql.SQLSyntaxErrorException: incompatible data type in conversion at org.hsqldb.jdbc.JDBCUtil.sqlException(Unknown Source) at org.hsqldb.jdbc.JDBCUtil.sqlException(Unknown Source) at org.hsqldb.jdbc.JDBCUtil.sqlException(Unknown Source) at org.hsqldb.jdbc.JDBCResultSet.getArray(Unknown Source) at org.jooq.tools.jdbc.DefaultResultSet.getArray(DefaultResultSet.java:688) at org.jooq.impl.CursorImpl$CursorResultSet.getArray(CursorImpl.java:437) at org.jooq.impl.DefaultBinding$DefaultArrayBinding.get0(DefaultBinding.java:1203) at org.jooq.impl.DefaultBinding$DefaultArrayBinding.get0(DefaultBinding.java:1) at org.jooq.impl.DefaultBinding$AbstractBinding.get(DefaultBinding.java:910) at org.jooq.impl.CursorImpl$CursorIterator$CursorRecordInitialiser.setValue(CursorImpl.java:1538) ... 13 more Caused by: org.hsqldb.HsqlException: incompatible data type in conversion at org.hsqldb.error.Error.error(Unknown Source) at org.hsqldb.error.Error.error(Unknown Source) ... 21 more ``` I've reported this as a bug: https://sourceforge.net/p/hsqldb/bugs/1620. Let's see if Fred Toussi agrees if it's a bug. If he doesn't, we'll work around this in jOOQ.
1.0
Cannot fetch inline NULL value as array type in HSQLDB - It seems we can't do this with HSQLDB: ```java ctx.select(inline(null, String[].class)).fetch(); ``` We're getting: ``` Exception in thread "main" org.jooq.exception.DataAccessException: SQL [select null from (values(1)) as dual(dual) -- SQL rendered with a free trial version of jOOQ 3.15.0-SNAPSHOT]; Error while reading field: null, at JDBC index: 1 at org.jooq_3.15.0-SNAPSHOT.HSQLDB.debug(Unknown Source) at org.jooq_3.15.0-SNAPSHOT.HSQLDB.debug(Unknown Source) at org.jooq.impl.Tools.translate(Tools.java:2877) at org.jooq.impl.DefaultExecuteContext.sqlException(DefaultExecuteContext.java:695) at org.jooq.impl.CursorImpl$CursorIterator.fetchNext(CursorImpl.java:1436) at org.jooq.impl.CursorImpl$CursorIterator.hasNext(CursorImpl.java:1397) at org.jooq.impl.CursorImpl.fetchNext(CursorImpl.java:217) at org.jooq.impl.AbstractCursor.fetch(AbstractCursor.java:182) at org.jooq.impl.AbstractCursor.fetch(AbstractCursor.java:93) at org.jooq.impl.AbstractResultQuery.execute(AbstractResultQuery.java:289) at org.jooq.impl.AbstractQuery.execute(AbstractQuery.java:374) at org.jooq.impl.AbstractResultQuery.fetch(AbstractResultQuery.java:314) at org.jooq.impl.SelectImpl.fetch(SelectImpl.java:2818) at org.jooq.testscripts.JDBC.main(JDBC.java:25) Caused by: java.sql.SQLException: Error while reading field: null, at JDBC index: 1 at org.jooq.impl.CursorImpl$CursorIterator$CursorRecordInitialiser.setValue(CursorImpl.java:1548) at org.jooq.impl.CursorImpl$CursorIterator$CursorRecordInitialiser.apply(CursorImpl.java:1497) at org.jooq.impl.CursorImpl$CursorIterator$CursorRecordInitialiser.apply(CursorImpl.java:1) at org.jooq.impl.RecordDelegate.operate(RecordDelegate.java:145) at org.jooq.impl.CursorImpl$CursorIterator.fetchNext(CursorImpl.java:1421) ... 9 more Caused by: java.sql.SQLSyntaxErrorException: incompatible data type in conversion at org.hsqldb.jdbc.JDBCUtil.sqlException(Unknown Source) at org.hsqldb.jdbc.JDBCUtil.sqlException(Unknown Source) at org.hsqldb.jdbc.JDBCUtil.sqlException(Unknown Source) at org.hsqldb.jdbc.JDBCResultSet.getArray(Unknown Source) at org.jooq.tools.jdbc.DefaultResultSet.getArray(DefaultResultSet.java:688) at org.jooq.impl.CursorImpl$CursorResultSet.getArray(CursorImpl.java:437) at org.jooq.impl.DefaultBinding$DefaultArrayBinding.get0(DefaultBinding.java:1203) at org.jooq.impl.DefaultBinding$DefaultArrayBinding.get0(DefaultBinding.java:1) at org.jooq.impl.DefaultBinding$AbstractBinding.get(DefaultBinding.java:910) at org.jooq.impl.CursorImpl$CursorIterator$CursorRecordInitialiser.setValue(CursorImpl.java:1538) ... 13 more Caused by: org.hsqldb.HsqlException: incompatible data type in conversion at org.hsqldb.error.Error.error(Unknown Source) at org.hsqldb.error.Error.error(Unknown Source) ... 21 more ``` I've reported this as a bug: https://sourceforge.net/p/hsqldb/bugs/1620. Let's see if Fred Toussi agrees if it's a bug. If he doesn't, we'll work around this in jOOQ.
defect
cannot fetch inline null value as array type in hsqldb it seems we can t do this with hsqldb java ctx select inline null string class fetch we re getting exception in thread main org jooq exception dataaccessexception sql error while reading field null at jdbc index at org jooq snapshot hsqldb debug unknown source at org jooq snapshot hsqldb debug unknown source at org jooq impl tools translate tools java at org jooq impl defaultexecutecontext sqlexception defaultexecutecontext java at org jooq impl cursorimpl cursoriterator fetchnext cursorimpl java at org jooq impl cursorimpl cursoriterator hasnext cursorimpl java at org jooq impl cursorimpl fetchnext cursorimpl java at org jooq impl abstractcursor fetch abstractcursor java at org jooq impl abstractcursor fetch abstractcursor java at org jooq impl abstractresultquery execute abstractresultquery java at org jooq impl abstractquery execute abstractquery java at org jooq impl abstractresultquery fetch abstractresultquery java at org jooq impl selectimpl fetch selectimpl java at org jooq testscripts jdbc main jdbc java caused by java sql sqlexception error while reading field null at jdbc index at org jooq impl cursorimpl cursoriterator cursorrecordinitialiser setvalue cursorimpl java at org jooq impl cursorimpl cursoriterator cursorrecordinitialiser apply cursorimpl java at org jooq impl cursorimpl cursoriterator cursorrecordinitialiser apply cursorimpl java at org jooq impl recorddelegate operate recorddelegate java at org jooq impl cursorimpl cursoriterator fetchnext cursorimpl java more caused by java sql sqlsyntaxerrorexception incompatible data type in conversion at org hsqldb jdbc jdbcutil sqlexception unknown source at org hsqldb jdbc jdbcutil sqlexception unknown source at org hsqldb jdbc jdbcutil sqlexception unknown source at org hsqldb jdbc jdbcresultset getarray unknown source at org jooq tools jdbc defaultresultset getarray defaultresultset java at org jooq impl cursorimpl cursorresultset getarray cursorimpl java at org jooq impl defaultbinding defaultarraybinding defaultbinding java at org jooq impl defaultbinding defaultarraybinding defaultbinding java at org jooq impl defaultbinding abstractbinding get defaultbinding java at org jooq impl cursorimpl cursoriterator cursorrecordinitialiser setvalue cursorimpl java more caused by org hsqldb hsqlexception incompatible data type in conversion at org hsqldb error error error unknown source at org hsqldb error error error unknown source more i ve reported this as a bug let s see if fred toussi agrees if it s a bug if he doesn t we ll work around this in jooq
1
13,943
8,743,485,262
IssuesEvent
2018-12-12 19:18:22
mercycorps/TolaActivity
https://api.github.com/repos/mercycorps/TolaActivity
closed
Add indicator form (full page version): Some validation messaging is missing
bug usability
**Note:** This problem is likely to occur frequently after Mangosteen because we are no longer auto-selecting a level for you. ## To reproduce the issue 1. On a program page in Demo, click "Add indicator". 2. On the Add indicator form, type in the indicator name and click "Save changes". 3. Go straight to the Target tab and fill in all required fields, choosing any target frequency except for LoP-only. 4. Click "Create targets". **Expected:** This message (screenshot from production), except it says Performance tab instead of Summary tab -- because you have not yet chosen a level. ![image](https://user-images.githubusercontent.com/33670923/49319592-60c9e400-f4b2-11e8-8d81-3df73570908d.png) **Observed:** I get no system feedback. # Acceptance criteria ## Scenario 1: I fail to select a level on the Performance tab 1. When I go through the steps described above and click "Create targets", I get the red error message, telling me to complete all required fields in the Performance tab. 2. As soon as I select a level, I can go back to the Target tab, click "Create targets" and get the results table as expected. ## Scenario 2: On the Summary tab, I purposely delete the indicator name and don't fill in a new one. 1. When I go through the steps described above and click "Create targets", I get this red error message. (This error validation does work as expected on production, but not on Demo.) ![image](https://user-images.githubusercontent.com/33670923/49319828-93281100-f4b3-11e8-8bb8-ad6835fe0b55.png) 2. As soon as I re-enter a name, I can go back to the Target tab, click "Create targets" and get the results table as expected. ## Scenario 3: I am extra naughty and don't select a level AND delete the name. Two messages should stack -- referencing the Summary and Performance tabs -- like they do in this example from production where I left fields empty on the Targets and Summary tab: ![image](https://user-images.githubusercontent.com/33670923/49319897-ea2de600-f4b3-11e8-964f-747c9ba433e3.png)
True
Add indicator form (full page version): Some validation messaging is missing - **Note:** This problem is likely to occur frequently after Mangosteen because we are no longer auto-selecting a level for you. ## To reproduce the issue 1. On a program page in Demo, click "Add indicator". 2. On the Add indicator form, type in the indicator name and click "Save changes". 3. Go straight to the Target tab and fill in all required fields, choosing any target frequency except for LoP-only. 4. Click "Create targets". **Expected:** This message (screenshot from production), except it says Performance tab instead of Summary tab -- because you have not yet chosen a level. ![image](https://user-images.githubusercontent.com/33670923/49319592-60c9e400-f4b2-11e8-8d81-3df73570908d.png) **Observed:** I get no system feedback. # Acceptance criteria ## Scenario 1: I fail to select a level on the Performance tab 1. When I go through the steps described above and click "Create targets", I get the red error message, telling me to complete all required fields in the Performance tab. 2. As soon as I select a level, I can go back to the Target tab, click "Create targets" and get the results table as expected. ## Scenario 2: On the Summary tab, I purposely delete the indicator name and don't fill in a new one. 1. When I go through the steps described above and click "Create targets", I get this red error message. (This error validation does work as expected on production, but not on Demo.) ![image](https://user-images.githubusercontent.com/33670923/49319828-93281100-f4b3-11e8-8bb8-ad6835fe0b55.png) 2. As soon as I re-enter a name, I can go back to the Target tab, click "Create targets" and get the results table as expected. ## Scenario 3: I am extra naughty and don't select a level AND delete the name. Two messages should stack -- referencing the Summary and Performance tabs -- like they do in this example from production where I left fields empty on the Targets and Summary tab: ![image](https://user-images.githubusercontent.com/33670923/49319897-ea2de600-f4b3-11e8-964f-747c9ba433e3.png)
non_defect
add indicator form full page version some validation messaging is missing note this problem is likely to occur frequently after mangosteen because we are no longer auto selecting a level for you to reproduce the issue on a program page in demo click add indicator on the add indicator form type in the indicator name and click save changes go straight to the target tab and fill in all required fields choosing any target frequency except for lop only click create targets expected this message screenshot from production except it says performance tab instead of summary tab because you have not yet chosen a level observed i get no system feedback acceptance criteria scenario i fail to select a level on the performance tab when i go through the steps described above and click create targets i get the red error message telling me to complete all required fields in the performance tab as soon as i select a level i can go back to the target tab click create targets and get the results table as expected scenario on the summary tab i purposely delete the indicator name and don t fill in a new one when i go through the steps described above and click create targets i get this red error message this error validation does work as expected on production but not on demo as soon as i re enter a name i can go back to the target tab click create targets and get the results table as expected scenario i am extra naughty and don t select a level and delete the name two messages should stack referencing the summary and performance tabs like they do in this example from production where i left fields empty on the targets and summary tab
0
128,073
12,359,313,981
IssuesEvent
2020-05-17 10:18:50
Tezzmo/PDS_Project
https://api.github.com/repos/Tezzmo/PDS_Project
opened
Guideline Python Notebook
description documentation
Create notebook as a guideline for all existing feature, where and when to execute them in terms of logic.
1.0
Guideline Python Notebook - Create notebook as a guideline for all existing feature, where and when to execute them in terms of logic.
non_defect
guideline python notebook create notebook as a guideline for all existing feature where and when to execute them in terms of logic
0
47,884
13,066,330,280
IssuesEvent
2020-07-30 21:28:07
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
closed
py2-v2 PYTHONPATH (Trac #1354)
Migrated from Trac cvmfs defect
When I source /cvmfs/icecube.opensciencegrid.org/py2-v2/RHEL_6_x86_64/metaprojects/icerec/trunk/env-shell.sh, my PYTHONPATH is set incorrectly and all python bindings cannot be found. It references '''/mnt/build'''/icecube.opensciencegrid.org/py2-v2/RHEL_6_x86_64/metaprojects/icerec/trunk/lib instead of /cvmfs/icecube.opensciencegrid.org/py2-v2/RHEL_6_x86_64/metaprojects/icerec/trunk/lib Migrated from https://code.icecube.wisc.edu/ticket/1354 ```json { "status": "closed", "changetime": "2015-09-22T19:35:21", "description": "When I source /cvmfs/icecube.opensciencegrid.org/py2-v2/RHEL_6_x86_64/metaprojects/icerec/trunk/env-shell.sh, my PYTHONPATH is set incorrectly and all python bindings cannot be found. \n\nIt references '''/mnt/build'''/icecube.opensciencegrid.org/py2-v2/RHEL_6_x86_64/metaprojects/icerec/trunk/lib instead of /cvmfs/icecube.opensciencegrid.org/py2-v2/RHEL_6_x86_64/metaprojects/icerec/trunk/lib\n\n", "reporter": "jkelley", "cc": "nega", "resolution": "fixed", "_ts": "1442950521994354", "component": "cvmfs", "summary": "py2-v2 PYTHONPATH", "priority": "major", "keywords": "", "time": "2015-09-17T23:28:15", "milestone": "", "owner": "david.schultz", "type": "defect" } ```
1.0
py2-v2 PYTHONPATH (Trac #1354) - When I source /cvmfs/icecube.opensciencegrid.org/py2-v2/RHEL_6_x86_64/metaprojects/icerec/trunk/env-shell.sh, my PYTHONPATH is set incorrectly and all python bindings cannot be found. It references '''/mnt/build'''/icecube.opensciencegrid.org/py2-v2/RHEL_6_x86_64/metaprojects/icerec/trunk/lib instead of /cvmfs/icecube.opensciencegrid.org/py2-v2/RHEL_6_x86_64/metaprojects/icerec/trunk/lib Migrated from https://code.icecube.wisc.edu/ticket/1354 ```json { "status": "closed", "changetime": "2015-09-22T19:35:21", "description": "When I source /cvmfs/icecube.opensciencegrid.org/py2-v2/RHEL_6_x86_64/metaprojects/icerec/trunk/env-shell.sh, my PYTHONPATH is set incorrectly and all python bindings cannot be found. \n\nIt references '''/mnt/build'''/icecube.opensciencegrid.org/py2-v2/RHEL_6_x86_64/metaprojects/icerec/trunk/lib instead of /cvmfs/icecube.opensciencegrid.org/py2-v2/RHEL_6_x86_64/metaprojects/icerec/trunk/lib\n\n", "reporter": "jkelley", "cc": "nega", "resolution": "fixed", "_ts": "1442950521994354", "component": "cvmfs", "summary": "py2-v2 PYTHONPATH", "priority": "major", "keywords": "", "time": "2015-09-17T23:28:15", "milestone": "", "owner": "david.schultz", "type": "defect" } ```
defect
pythonpath trac when i source cvmfs icecube opensciencegrid org rhel metaprojects icerec trunk env shell sh my pythonpath is set incorrectly and all python bindings cannot be found it references mnt build icecube opensciencegrid org rhel metaprojects icerec trunk lib instead of cvmfs icecube opensciencegrid org rhel metaprojects icerec trunk lib migrated from json status closed changetime description when i source cvmfs icecube opensciencegrid org rhel metaprojects icerec trunk env shell sh my pythonpath is set incorrectly and all python bindings cannot be found n nit references mnt build icecube opensciencegrid org rhel metaprojects icerec trunk lib instead of cvmfs icecube opensciencegrid org rhel metaprojects icerec trunk lib n n reporter jkelley cc nega resolution fixed ts component cvmfs summary pythonpath priority major keywords time milestone owner david schultz type defect
1
34,410
7,451,168,002
IssuesEvent
2018-03-29 01:17:11
kerdokullamae/test_koik_issued
https://api.github.com/repos/kerdokullamae/test_koik_issued
closed
Uue KY ja isiku lisamine
P: highest R: fixed T: defect
**Reported by jelenag on 11 Mar 2013 12:18 UTC** Uue KY salvestamisel sain veateade: ``` Unable to execute INSERT statement [INTO dira.search_index_sync_queue (ID, OBJECT_TYPE_ID, OBJECT_ID, ACTION_ID) VALUES (:p0, :p1, :p2, :p3)](INSERT) [SQLSTATE[23505](wrapped:): Unique violation: 7 ERROR: duplicate key value violates unique constraint "search_index_sync_queue_uq" DETAIL: Key (object_type_id, object_id)=(DESCRIPTION_UNIT, 117271) already exists.] ``` Uue isiku salvestamisel: ``` Unable to execute INSERT statement [INTO dira.search_index_sync_queue (ID, OBJECT_TYPE_ID, OBJECT_ID, ACTION_ID) VALUES (:p0, :p1, :p2, :p3)](INSERT) [SQLSTATE[23505](wrapped:): Unique violation: 7 ERROR: duplicate key value violates unique constraint "search_index_sync_queue_uq" DETAIL: Key (object_type_id, object_id)=(PERSON, 48) already exists.] ```
1.0
Uue KY ja isiku lisamine - **Reported by jelenag on 11 Mar 2013 12:18 UTC** Uue KY salvestamisel sain veateade: ``` Unable to execute INSERT statement [INTO dira.search_index_sync_queue (ID, OBJECT_TYPE_ID, OBJECT_ID, ACTION_ID) VALUES (:p0, :p1, :p2, :p3)](INSERT) [SQLSTATE[23505](wrapped:): Unique violation: 7 ERROR: duplicate key value violates unique constraint "search_index_sync_queue_uq" DETAIL: Key (object_type_id, object_id)=(DESCRIPTION_UNIT, 117271) already exists.] ``` Uue isiku salvestamisel: ``` Unable to execute INSERT statement [INTO dira.search_index_sync_queue (ID, OBJECT_TYPE_ID, OBJECT_ID, ACTION_ID) VALUES (:p0, :p1, :p2, :p3)](INSERT) [SQLSTATE[23505](wrapped:): Unique violation: 7 ERROR: duplicate key value violates unique constraint "search_index_sync_queue_uq" DETAIL: Key (object_type_id, object_id)=(PERSON, 48) already exists.] ```
defect
uue ky ja isiku lisamine reported by jelenag on mar utc uue ky salvestamisel sain veateade unable to execute insert statement insert wrapped unique violation error duplicate key value violates unique constraint search index sync queue uq detail key object type id object id description unit already exists uue isiku salvestamisel unable to execute insert statement insert wrapped unique violation error duplicate key value violates unique constraint search index sync queue uq detail key object type id object id person already exists
1
312,900
26,884,997,782
IssuesEvent
2023-02-06 01:51:10
metafizzy/flickity
https://api.github.com/repos/metafizzy/flickity
closed
carousel no working properly when using *ngfor to display dynamic data
test case required
Flickity is a very good carousel, but I'm having an issue using it in angular when using static data it works good (sometimes it's not showing properly and after reloading it works) but when using a dynamic data returned from a service it shows all slides and images as a one block with no slides how to fix this issue ? code i'm using ``` <div class="carousel cssanimation sequence fadeInBottom" data-flickity='{"autoPlay": "10000", "wrapAround": true, "imagesLoaded": true}'> <a href="" *ngFor="let show of RecentReleases"> <img class="carousel-cell" [src]="show.largePoster" alt="poster"> <div class="movie-info d-flex flex-column align-items-start"> <p class="title">{{show.title}}</p> <div class="genres d-flex flex-row" *ngFor="let genre of show.genres"> <p class="genre">{{genre}}</p> <p class="separator">|</p> </div> </div> </a> </div> ``` and my ts code ``` RecentReleases?: IShow[]; ngOnInit(): void { this.homeService.getRecentReleases().subscribe((shows) => { this.RecentReleases = shows; }); } ```
1.0
carousel no working properly when using *ngfor to display dynamic data - Flickity is a very good carousel, but I'm having an issue using it in angular when using static data it works good (sometimes it's not showing properly and after reloading it works) but when using a dynamic data returned from a service it shows all slides and images as a one block with no slides how to fix this issue ? code i'm using ``` <div class="carousel cssanimation sequence fadeInBottom" data-flickity='{"autoPlay": "10000", "wrapAround": true, "imagesLoaded": true}'> <a href="" *ngFor="let show of RecentReleases"> <img class="carousel-cell" [src]="show.largePoster" alt="poster"> <div class="movie-info d-flex flex-column align-items-start"> <p class="title">{{show.title}}</p> <div class="genres d-flex flex-row" *ngFor="let genre of show.genres"> <p class="genre">{{genre}}</p> <p class="separator">|</p> </div> </div> </a> </div> ``` and my ts code ``` RecentReleases?: IShow[]; ngOnInit(): void { this.homeService.getRecentReleases().subscribe((shows) => { this.RecentReleases = shows; }); } ```
non_defect
carousel no working properly when using ngfor to display dynamic data flickity is a very good carousel but i m having an issue using it in angular when using static data it works good sometimes it s not showing properly and after reloading it works but when using a dynamic data returned from a service it shows all slides and images as a one block with no slides how to fix this issue code i m using div class carousel cssanimation sequence fadeinbottom data flickity autoplay wraparound true imagesloaded true show title genre and my ts code recentreleases ishow ngoninit void this homeservice getrecentreleases subscribe shows this recentreleases shows
0
305,672
26,401,375,295
IssuesEvent
2023-01-13 01:42:20
epicmaxco/vuestic-ui
https://api.github.com/repos/epicmaxco/vuestic-ui
closed
Create Toggle between dark and light theme for vue-book
LOW PRIORITY test
We need to check components on a dark and light backgrounds.
1.0
Create Toggle between dark and light theme for vue-book - We need to check components on a dark and light backgrounds.
non_defect
create toggle between dark and light theme for vue book we need to check components on a dark and light backgrounds
0
9,642
2,615,164,297
IssuesEvent
2015-03-01 06:44:15
chrsmith/reaver-wps
https://api.github.com/repos/chrsmith/reaver-wps
opened
reaver : command not found even when everything i need like reaver and libsqlite is installed
auto-migrated Priority-Triage Type-Defect
``` A few things to consider before submitting an issue: 0. We write documentation for a reason, if you have not read it and are having problems with Reaver these pages are required reading before submitting an issue: http://code.google.com/p/reaver-wps/wiki/HintsAndTips http://code.google.com/p/reaver-wps/wiki/README http://code.google.com/p/reaver-wps/wiki/FAQ http://code.google.com/p/reaver-wps/wiki/SupportedWirelessDrivers 1. Reaver will only work if your card is in monitor mode. If you do not know what monitor mode is then you should learn more about 802.11 hacking in linux before using Reaver. 2. Using Reaver against access points you do not own or have permission to attack is illegal. If you cannot answer basic questions (i.e. model number, distance away, etc) about the device you are attacking then do not post your issue here. We will not help you break the law. 3. Please look through issues that have already been posted and make sure your question has not already been asked here: http://code.google.com/p /reaver-wps/issues/list 4. Often times we need packet captures of mon0 while Reaver is running to troubleshoot the issue (tcpdump -i mon0 -s0 -w broken_reaver.pcap). Issue reports with pcap files attached will receive more serious consideration. Answer the following questions for every issue submitted: 0. What version of Reaver are you using? (Only defects against the latest version will be considered.) 1.4 1. What operating system are you using (Linux is the only supported OS)? Ubuntu 2. Is your wireless card in monitor mode (yes/no)? yes 3. What is the signal strength of the Access Point you are trying to crack? -73 4. What is the manufacturer and model # of the device you are trying to crack? Netgear MODEL NR. UNKNOWN 5. What is the entire command line string you are supplying to reaver? sudo reaver -i mon0 -b <BSSID> -vv 6. Please describe what you think the issue is. no clue at all. 7. Paste the output from Reaver below. reaver: command not found. ``` Original issue reported on code.google.com by `Cristian...@gmail.com` on 25 Sep 2012 at 11:46
1.0
reaver : command not found even when everything i need like reaver and libsqlite is installed - ``` A few things to consider before submitting an issue: 0. We write documentation for a reason, if you have not read it and are having problems with Reaver these pages are required reading before submitting an issue: http://code.google.com/p/reaver-wps/wiki/HintsAndTips http://code.google.com/p/reaver-wps/wiki/README http://code.google.com/p/reaver-wps/wiki/FAQ http://code.google.com/p/reaver-wps/wiki/SupportedWirelessDrivers 1. Reaver will only work if your card is in monitor mode. If you do not know what monitor mode is then you should learn more about 802.11 hacking in linux before using Reaver. 2. Using Reaver against access points you do not own or have permission to attack is illegal. If you cannot answer basic questions (i.e. model number, distance away, etc) about the device you are attacking then do not post your issue here. We will not help you break the law. 3. Please look through issues that have already been posted and make sure your question has not already been asked here: http://code.google.com/p /reaver-wps/issues/list 4. Often times we need packet captures of mon0 while Reaver is running to troubleshoot the issue (tcpdump -i mon0 -s0 -w broken_reaver.pcap). Issue reports with pcap files attached will receive more serious consideration. Answer the following questions for every issue submitted: 0. What version of Reaver are you using? (Only defects against the latest version will be considered.) 1.4 1. What operating system are you using (Linux is the only supported OS)? Ubuntu 2. Is your wireless card in monitor mode (yes/no)? yes 3. What is the signal strength of the Access Point you are trying to crack? -73 4. What is the manufacturer and model # of the device you are trying to crack? Netgear MODEL NR. UNKNOWN 5. What is the entire command line string you are supplying to reaver? sudo reaver -i mon0 -b <BSSID> -vv 6. Please describe what you think the issue is. no clue at all. 7. Paste the output from Reaver below. reaver: command not found. ``` Original issue reported on code.google.com by `Cristian...@gmail.com` on 25 Sep 2012 at 11:46
defect
reaver command not found even when everything i need like reaver and libsqlite is installed a few things to consider before submitting an issue we write documentation for a reason if you have not read it and are having problems with reaver these pages are required reading before submitting an issue reaver will only work if your card is in monitor mode if you do not know what monitor mode is then you should learn more about hacking in linux before using reaver using reaver against access points you do not own or have permission to attack is illegal if you cannot answer basic questions i e model number distance away etc about the device you are attacking then do not post your issue here we will not help you break the law please look through issues that have already been posted and make sure your question has not already been asked here reaver wps issues list often times we need packet captures of while reaver is running to troubleshoot the issue tcpdump i w broken reaver pcap issue reports with pcap files attached will receive more serious consideration answer the following questions for every issue submitted what version of reaver are you using only defects against the latest version will be considered what operating system are you using linux is the only supported os ubuntu is your wireless card in monitor mode yes no yes what is the signal strength of the access point you are trying to crack what is the manufacturer and model of the device you are trying to crack netgear model nr unknown what is the entire command line string you are supplying to reaver sudo reaver i b vv please describe what you think the issue is no clue at all paste the output from reaver below reaver command not found original issue reported on code google com by cristian gmail com on sep at
1
54,661
13,820,658,547
IssuesEvent
2020-10-13 00:09:04
google/guava
https://api.github.com/repos/google/guava
closed
Slicing a ByteSource twice throws an unexpected IllegalArgumentException
P3 package=io status=duplicate type=defect
Currently, trying to slice a `ByteSource` twice throws an unexpected `IllegalArgumentException` if the `ByteSource` returned after the first slice has a length smaller than the offset of the subsequent slice. For example, this code throws an `IllegalArgumentException`: ``` import com.google.common.io.ByteSource; public class GuavaTest { public static void main(String[] args) { ByteSource.concat().slice(0, 3).slice(4, 3); } } ``` That is against the documentation, where it is specified that it should return an empty `ByteSource`.
1.0
Slicing a ByteSource twice throws an unexpected IllegalArgumentException - Currently, trying to slice a `ByteSource` twice throws an unexpected `IllegalArgumentException` if the `ByteSource` returned after the first slice has a length smaller than the offset of the subsequent slice. For example, this code throws an `IllegalArgumentException`: ``` import com.google.common.io.ByteSource; public class GuavaTest { public static void main(String[] args) { ByteSource.concat().slice(0, 3).slice(4, 3); } } ``` That is against the documentation, where it is specified that it should return an empty `ByteSource`.
defect
slicing a bytesource twice throws an unexpected illegalargumentexception currently trying to slice a bytesource twice throws an unexpected illegalargumentexception if the bytesource returned after the first slice has a length smaller than the offset of the subsequent slice for example this code throws an illegalargumentexception import com google common io bytesource public class guavatest public static void main string args bytesource concat slice slice that is against the documentation where it is specified that it should return an empty bytesource
1
24,262
3,941,468,811
IssuesEvent
2016-04-27 07:52:49
JulienGenoud/renderscript-examples
https://api.github.com/repos/JulienGenoud/renderscript-examples
closed
Page Curl is not working in 4.1
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Downloaded pagecurl project 2. Imported to Eclipse ->API level 16 3. Created AVD with option GPU on 4. Run the application What is the expected output? What do you see instead? Expected Should show the Pagecurl but Actual No rendering,Black screen What version of the product are you using? On what operating system? Windows 7 32bit,Android Emulator Please provide any additional information below. ``` Original issue reported on code.google.com by `kuvetha...@gmail.com` on 17 Jul 2012 at 6:17
1.0
Page Curl is not working in 4.1 - ``` What steps will reproduce the problem? 1. Downloaded pagecurl project 2. Imported to Eclipse ->API level 16 3. Created AVD with option GPU on 4. Run the application What is the expected output? What do you see instead? Expected Should show the Pagecurl but Actual No rendering,Black screen What version of the product are you using? On what operating system? Windows 7 32bit,Android Emulator Please provide any additional information below. ``` Original issue reported on code.google.com by `kuvetha...@gmail.com` on 17 Jul 2012 at 6:17
defect
page curl is not working in what steps will reproduce the problem downloaded pagecurl project imported to eclipse api level created avd with option gpu on run the application what is the expected output what do you see instead expected should show the pagecurl but actual no rendering black screen what version of the product are you using on what operating system windows android emulator please provide any additional information below original issue reported on code google com by kuvetha gmail com on jul at
1
112,827
17,104,047,712
IssuesEvent
2021-07-09 15:05:14
brogers588/Java_Demo
https://api.github.com/repos/brogers588/Java_Demo
opened
CVE-2016-2510 (High) detected in bsh-core-2.0b4.jar
security vulnerability
## CVE-2016-2510 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bsh-core-2.0b4.jar</b></p></summary> <p>BeanShell core</p> <p>Path to dependency file: Java_Demo/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/beanshell/bsh-core/2.0b4/bsh-core-2.0b4.jar</p> <p> Dependency Hierarchy: - esapi-2.1.0.1.jar (Root Library) - :x: **bsh-core-2.0b4.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/brogers588/Java_Demo/commit/057e2c009e307c82b86a18912147951b456bf408">057e2c009e307c82b86a18912147951b456bf408</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> BeanShell (bsh) before 2.0b6, when included on the classpath by an application that uses Java serialization or XStream, allows remote attackers to execute arbitrary code via crafted serialized data, related to XThis.Handler. <p>Publish Date: 2016-04-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-2510>CVE-2016-2510</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-2510">https://nvd.nist.gov/vuln/detail/CVE-2016-2510</a></p> <p>Release Date: 2016-04-07</p> <p>Fix Resolution: 2.0b6</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.beanshell","packageName":"bsh-core","packageVersion":"2.0b4","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.owasp.esapi:esapi:2.1.0.1;org.beanshell:bsh-core:2.0b4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.0b6"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2016-2510","vulnerabilityDetails":"BeanShell (bsh) before 2.0b6, when included on the classpath by an application that uses Java serialization or XStream, allows remote attackers to execute arbitrary code via crafted serialized data, related to XThis.Handler.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-2510","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2016-2510 (High) detected in bsh-core-2.0b4.jar - ## CVE-2016-2510 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bsh-core-2.0b4.jar</b></p></summary> <p>BeanShell core</p> <p>Path to dependency file: Java_Demo/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/beanshell/bsh-core/2.0b4/bsh-core-2.0b4.jar</p> <p> Dependency Hierarchy: - esapi-2.1.0.1.jar (Root Library) - :x: **bsh-core-2.0b4.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/brogers588/Java_Demo/commit/057e2c009e307c82b86a18912147951b456bf408">057e2c009e307c82b86a18912147951b456bf408</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> BeanShell (bsh) before 2.0b6, when included on the classpath by an application that uses Java serialization or XStream, allows remote attackers to execute arbitrary code via crafted serialized data, related to XThis.Handler. <p>Publish Date: 2016-04-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-2510>CVE-2016-2510</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-2510">https://nvd.nist.gov/vuln/detail/CVE-2016-2510</a></p> <p>Release Date: 2016-04-07</p> <p>Fix Resolution: 2.0b6</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.beanshell","packageName":"bsh-core","packageVersion":"2.0b4","packageFilePaths":["/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.owasp.esapi:esapi:2.1.0.1;org.beanshell:bsh-core:2.0b4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.0b6"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2016-2510","vulnerabilityDetails":"BeanShell (bsh) before 2.0b6, when included on the classpath by an application that uses Java serialization or XStream, allows remote attackers to execute arbitrary code via crafted serialized data, related to XThis.Handler.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-2510","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_defect
cve high detected in bsh core jar cve high severity vulnerability vulnerable library bsh core jar beanshell core path to dependency file java demo pom xml path to vulnerable library home wss scanner repository org beanshell bsh core bsh core jar dependency hierarchy esapi jar root library x bsh core jar vulnerable library found in head commit a href found in base branch master vulnerability details beanshell bsh before when included on the classpath by an application that uses java serialization or xstream allows remote attackers to execute arbitrary code via crafted serialized data related to xthis handler publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org owasp esapi esapi org beanshell bsh core isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails beanshell bsh before when included on the classpath by an application that uses java serialization or xstream allows remote attackers to execute arbitrary code via crafted serialized data related to xthis handler vulnerabilityurl
0
23,207
3,776,303,141
IssuesEvent
2016-03-17 16:15:02
obophenotype/upheno
https://api.github.com/repos/obophenotype/upheno
closed
term replaced by issues
Priority-Medium Status-Accepted Type-Defect
Originally reported on Google Code with ID 112 ``` In the Monarch import chain in SciGraph, there are different ranges for the 'term replaced by' property. For example, some use a string: <obo:IAO_0100001 rdf:datatype="http://www.w3.org/2001/XMLSchema#string">GO:0050801</obo:IAO_0100001> Some use an IRI: <obo:IAO_0100001 rdf:resource="http://purl.obolibrary.org/obo/HP_0008479"/> Chris C. kindly generated me a report of what is in Monarch.owl now, attached. The first set of results (927 in total) are just mapped as XMLSchema#string The latter 17 results are mapped as IRIs. Ideally we could convert the former to the latter and use the IRIs, or have some functionality to realize these strings as nodes in SciGraph. There are MP, HP, UBERON, SO terms, etc in there. ``` Reported by `haendel@ohsu.edu` on 2015-03-03 01:41:39 <hr> * *Attachment: [IAO_0100001_usage.txt](https://storage.googleapis.com/google-code-attachments/phenotype-ontologies/issue-112/comment-0/IAO_0100001_usage.txt)*
1.0
term replaced by issues - Originally reported on Google Code with ID 112 ``` In the Monarch import chain in SciGraph, there are different ranges for the 'term replaced by' property. For example, some use a string: <obo:IAO_0100001 rdf:datatype="http://www.w3.org/2001/XMLSchema#string">GO:0050801</obo:IAO_0100001> Some use an IRI: <obo:IAO_0100001 rdf:resource="http://purl.obolibrary.org/obo/HP_0008479"/> Chris C. kindly generated me a report of what is in Monarch.owl now, attached. The first set of results (927 in total) are just mapped as XMLSchema#string The latter 17 results are mapped as IRIs. Ideally we could convert the former to the latter and use the IRIs, or have some functionality to realize these strings as nodes in SciGraph. There are MP, HP, UBERON, SO terms, etc in there. ``` Reported by `haendel@ohsu.edu` on 2015-03-03 01:41:39 <hr> * *Attachment: [IAO_0100001_usage.txt](https://storage.googleapis.com/google-code-attachments/phenotype-ontologies/issue-112/comment-0/IAO_0100001_usage.txt)*
defect
term replaced by issues originally reported on google code with id in the monarch import chain in scigraph there are different ranges for the term replaced by property for example some use a string obo iao rdf datatype some use an iri obo iao rdf resource chris c kindly generated me a report of what is in monarch owl now attached the first set of results in total are just mapped as xmlschema string the latter results are mapped as iris ideally we could convert the former to the latter and use the iris or have some functionality to realize these strings as nodes in scigraph there are mp hp uberon so terms etc in there reported by haendel ohsu edu on attachment
1
67,057
20,826,318,492
IssuesEvent
2022-03-18 21:26:36
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Space room list header covers the list
T-Defect S-Minor A-Spaces-Settings A-Appearance O-Occasional
### Steps to reproduce 1. Open Space setting 2. Add some rooms 3. Change window size ### Outcome #### What did you expect? The list header should be wrapped. https://user-images.githubusercontent.com/3362943/158959733-d14230e6-fbb4-4cae-830d-311883d818e6.mp4 #### What happened instead? The buttons covers the room list, and you cannot select a room. https://user-images.githubusercontent.com/3362943/158959769-a64df65d-f608-45c0-b349-abdf253736c9.mp4 ### Operating system _No response_ ### Browser information Firefox ### URL for webapp localhost ### Application version develop branch ### Homeserver _No response_ ### Will you send logs? No
1.0
Space room list header covers the list - ### Steps to reproduce 1. Open Space setting 2. Add some rooms 3. Change window size ### Outcome #### What did you expect? The list header should be wrapped. https://user-images.githubusercontent.com/3362943/158959733-d14230e6-fbb4-4cae-830d-311883d818e6.mp4 #### What happened instead? The buttons covers the room list, and you cannot select a room. https://user-images.githubusercontent.com/3362943/158959769-a64df65d-f608-45c0-b349-abdf253736c9.mp4 ### Operating system _No response_ ### Browser information Firefox ### URL for webapp localhost ### Application version develop branch ### Homeserver _No response_ ### Will you send logs? No
defect
space room list header covers the list steps to reproduce open space setting add some rooms change window size outcome what did you expect the list header should be wrapped what happened instead the buttons covers the room list and you cannot select a room operating system no response browser information firefox url for webapp localhost application version develop branch homeserver no response will you send logs no
1