Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
75,619
3,470,203,071
IssuesEvent
2015-12-23 05:51:51
WebDevJL/EasyMvc
https://api.github.com/repos/WebDevJL/EasyMvc
closed
Generate any kind of class given the minimum necessary info
priority:high
Enable to specific: - Class description => ok - Class name =>ok - Class derivation => - Class interface to implement => - Class destination folder (will be used to calculate the namespace) => ok - Class properties with generation of get/set if necessary => later - Class methods with name, parameters (using type-hinting) and default return type => later NB: If return type is an object, the value provided will be the namespace of the Class to instanciate.
1.0
Generate any kind of class given the minimum necessary info - Enable to specific: - Class description => ok - Class name =>ok - Class derivation => - Class interface to implement => - Class destination folder (will be used to calculate the namespace) => ok - Class properties with generation of get/set if necessary => later - Class methods with name, parameters (using type-hinting) and default return type => later NB: If return type is an object, the value provided will be the namespace of the Class to instanciate.
non_process
generate any kind of class given the minimum necessary info enable to specific class description ok class name ok class derivation class interface to implement class destination folder will be used to calculate the namespace ok class properties with generation of get set if necessary later class methods with name parameters using type hinting and default return type later nb if return type is an object the value provided will be the namespace of the class to instanciate
0
9,283
12,303,889,802
IssuesEvent
2020-05-11 19:30:20
nextgenhealthcare/connect
https://api.github.com/repos/nextgenhealthcare/connect
closed
Allow users to return a Response from the Postprocessor script
post postprocessor processor response responseMap responsemap.put
Instead of making users call responseMap.put to place response map variables (and then having to select that variable in the source connector panel), it would be helpful to allow users to choose "Postprocessor" in the response combo box, and then simply return a response from the postprocessor. This would be the new Donkey Response object, so the ResponseFactory and reference list would need to be updated as well. Imported Issue. Original Details: Jira Issue Key: MIRTH-2279 Reporter: narupley Created: 2012-11-28T13:52:00.000-0800
2.0
Allow users to return a Response from the Postprocessor script - Instead of making users call responseMap.put to place response map variables (and then having to select that variable in the source connector panel), it would be helpful to allow users to choose "Postprocessor" in the response combo box, and then simply return a response from the postprocessor. This would be the new Donkey Response object, so the ResponseFactory and reference list would need to be updated as well. Imported Issue. Original Details: Jira Issue Key: MIRTH-2279 Reporter: narupley Created: 2012-11-28T13:52:00.000-0800
process
allow users to return a response from the postprocessor script instead of making users call responsemap put to place response map variables and then having to select that variable in the source connector panel it would be helpful to allow users to choose postprocessor in the response combo box and then simply return a response from the postprocessor this would be the new donkey response object so the responsefactory and reference list would need to be updated as well imported issue original details jira issue key mirth reporter narupley created
1
15,360
19,531,218,492
IssuesEvent
2021-12-30 17:11:39
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
multiprocessing ProcessException (and subclasses) can't be pickled/unpickled
module: multiprocessing triaged
### 🐛 Describe the bug Error instances that inherit from [`ProcessException`](https://github.com/pytorch/pytorch/blob/1065739781cae67f7861beaaceadb736e5e52271/torch/multiprocessing/spawn.py#L12) cannot be picked/unpickled. For example: ```python >>> import dill >>> from torch.multiprocessing import ProcessRaisedException >>> e = ProcessRaisedException("Oh no!", 1, 1) >>> dill.loads(dill.dumps(e)) ``` ``` Traceback (most recent call last): dill.loads(dill.dumps(e)) File ".../lib/python3.9/site-packages/dill/_dill.py", line 327, in loads return load(file, ignore, **kwds) File ".../lib/python3.9/site-packages/dill/_dill.py", line 313, in load return Unpickler(file, ignore=ignore, **kwds).load() File ".../lib/python3.9/site-packages/dill/_dill.py", line 525, in load obj = StockUnpickler.load(self) TypeError: __init__() missing 2 required positional arguments: 'error_index' and 'error_pid' ``` This is easy to fix and I wouldn't mind submitting a PR to do so. We'd just need to override `__reduce__` on `ProcessException`: ```python class ProcessException(Exception): __slots__ = ["error_index", "error_pid"] def __init__(self, msg: str, error_index: int, pid: int): super().__init__(msg) self.error_index = error_index self.pid = pid def __reduce__(self): return (self.__class__, self.args + (self.error_index, self.pid), {}) ``` ### Versions Collecting environment information... PyTorch version: 1.10.1+cu113 Is debug build: False CUDA used to build PyTorch: 11.3 ROCM used to build PyTorch: N/A OS: Ubuntu 20.04.3 LTS (x86_64) GCC version: (Ubuntu 9.3.0-17ubuntu1~20.04) 9.3.0 Clang version: Could not collect CMake version: version 3.16.3 Libc version: glibc-2.31 Python version: 3.9.7 (default, Sep 9 2021, 23:20:13) [GCC 9.3.0] (64-bit runtime) Python platform: Linux-5.11.0-1023-gcp-x86_64-with-glibc2.31 Is CUDA available: True CUDA runtime version: Could not collect GPU models and configuration: GPU 0: NVIDIA A100-SXM4-40GB GPU 1: NVIDIA A100-SXM4-40GB Nvidia driver version: 470.86 cuDNN version: Could not collect HIP runtime version: N/A MIOpen runtime version: N/A Versions of relevant libraries: [pip3] mypy==0.910 [pip3] mypy-extensions==0.4.3 [pip3] numpy==1.21.2 [pip3] torch==1.10.1+cu113 [conda] Could not collect cc @VitalyFedyunin
1.0
multiprocessing ProcessException (and subclasses) can't be pickled/unpickled - ### 🐛 Describe the bug Error instances that inherit from [`ProcessException`](https://github.com/pytorch/pytorch/blob/1065739781cae67f7861beaaceadb736e5e52271/torch/multiprocessing/spawn.py#L12) cannot be picked/unpickled. For example: ```python >>> import dill >>> from torch.multiprocessing import ProcessRaisedException >>> e = ProcessRaisedException("Oh no!", 1, 1) >>> dill.loads(dill.dumps(e)) ``` ``` Traceback (most recent call last): dill.loads(dill.dumps(e)) File ".../lib/python3.9/site-packages/dill/_dill.py", line 327, in loads return load(file, ignore, **kwds) File ".../lib/python3.9/site-packages/dill/_dill.py", line 313, in load return Unpickler(file, ignore=ignore, **kwds).load() File ".../lib/python3.9/site-packages/dill/_dill.py", line 525, in load obj = StockUnpickler.load(self) TypeError: __init__() missing 2 required positional arguments: 'error_index' and 'error_pid' ``` This is easy to fix and I wouldn't mind submitting a PR to do so. We'd just need to override `__reduce__` on `ProcessException`: ```python class ProcessException(Exception): __slots__ = ["error_index", "error_pid"] def __init__(self, msg: str, error_index: int, pid: int): super().__init__(msg) self.error_index = error_index self.pid = pid def __reduce__(self): return (self.__class__, self.args + (self.error_index, self.pid), {}) ``` ### Versions Collecting environment information... PyTorch version: 1.10.1+cu113 Is debug build: False CUDA used to build PyTorch: 11.3 ROCM used to build PyTorch: N/A OS: Ubuntu 20.04.3 LTS (x86_64) GCC version: (Ubuntu 9.3.0-17ubuntu1~20.04) 9.3.0 Clang version: Could not collect CMake version: version 3.16.3 Libc version: glibc-2.31 Python version: 3.9.7 (default, Sep 9 2021, 23:20:13) [GCC 9.3.0] (64-bit runtime) Python platform: Linux-5.11.0-1023-gcp-x86_64-with-glibc2.31 Is CUDA available: True CUDA runtime version: Could not collect GPU models and configuration: GPU 0: NVIDIA A100-SXM4-40GB GPU 1: NVIDIA A100-SXM4-40GB Nvidia driver version: 470.86 cuDNN version: Could not collect HIP runtime version: N/A MIOpen runtime version: N/A Versions of relevant libraries: [pip3] mypy==0.910 [pip3] mypy-extensions==0.4.3 [pip3] numpy==1.21.2 [pip3] torch==1.10.1+cu113 [conda] Could not collect cc @VitalyFedyunin
process
multiprocessing processexception and subclasses can t be pickled unpickled 🐛 describe the bug error instances that inherit from cannot be picked unpickled for example python import dill from torch multiprocessing import processraisedexception e processraisedexception oh no dill loads dill dumps e traceback most recent call last dill loads dill dumps e file lib site packages dill dill py line in loads return load file ignore kwds file lib site packages dill dill py line in load return unpickler file ignore ignore kwds load file lib site packages dill dill py line in load obj stockunpickler load self typeerror init missing required positional arguments error index and error pid this is easy to fix and i wouldn t mind submitting a pr to do so we d just need to override reduce on processexception python class processexception exception slots def init self msg str error index int pid int super init msg self error index error index self pid pid def reduce self return self class self args self error index self pid versions collecting environment information pytorch version is debug build false cuda used to build pytorch rocm used to build pytorch n a os ubuntu lts gcc version ubuntu clang version could not collect cmake version version libc version glibc python version default sep bit runtime python platform linux gcp with is cuda available true cuda runtime version could not collect gpu models and configuration gpu nvidia gpu nvidia nvidia driver version cudnn version could not collect hip runtime version n a miopen runtime version n a versions of relevant libraries mypy mypy extensions numpy torch could not collect cc vitalyfedyunin
1
15,651
19,846,744,120
IssuesEvent
2022-01-21 07:34:21
ooi-data/CE04OSSM-RID27-02-FLORTD000-recovered_host-flort_sample
https://api.github.com/repos/ooi-data/CE04OSSM-RID27-02-FLORTD000-recovered_host-flort_sample
opened
🛑 Processing failed: ValueError
process
## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T07:34:20.963174. ## Details Flow name: `CE04OSSM-RID27-02-FLORTD000-recovered_host-flort_sample` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
1.0
🛑 Processing failed: ValueError - ## Overview `ValueError` found in `processing_task` task during run ended on 2022-01-21T07:34:20.963174. ## Details Flow name: `CE04OSSM-RID27-02-FLORTD000-recovered_host-flort_sample` Task name: `processing_task` Error type: `ValueError` Error message: not enough values to unpack (expected 3, got 0) <details> <summary>Traceback</summary> ``` Traceback (most recent call last): File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing final_path = finalize_data_stream( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream append_to_zarr(mod_ds, final_store, enc, logger=logger) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr _append_zarr(store, mod_ds) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr existing_arr.append(var_data.values) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values return _as_array_or_item(self._data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item data = np.asarray(data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__ x = self.compute() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute (result,) = compute(self, traverse=False, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute results = schedule(dsk, keys, **kwargs) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get results = get_async( File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async raise_exception(exc, tb) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise raise exc File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task result = _execute_task(task, data) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task return func(*(_execute_task(a, cache) for a in args)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter c = np.asarray(c) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__ self._ensure_cached() File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached self.array = NumpyIndexingAdapter(np.asarray(self.array)) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__ return np.asarray(self.array, dtype=dtype) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__ return np.asarray(array[self.key], dtype=None) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__ return array[key.tuple] File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__ return self.get_basic_selection(selection, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection return self._get_basic_selection_nd(selection=selection, out=out, File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd return self._get_selection(indexer=indexer, out=out, fields=fields) File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection lchunk_coords, lchunk_selection, lout_selection = zip(*indexer) ValueError: not enough values to unpack (expected 3, got 0) ``` </details>
process
🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name recovered host flort sample task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got
1
68,422
8,287,882,212
IssuesEvent
2018-09-19 10:09:34
cosmos/voyager
https://api.github.com/repos/cosmos/voyager
opened
Change PageValidator boxes background color
design staking1
Description: <!-- Steps to reproduce, logs, and screenshots are helpful for us to resolve the bug --> As per our discussion in #1317. The current boxes are similar to our input form fields. Let's change the background color to make it more distinguishable.
1.0
Change PageValidator boxes background color - Description: <!-- Steps to reproduce, logs, and screenshots are helpful for us to resolve the bug --> As per our discussion in #1317. The current boxes are similar to our input form fields. Let's change the background color to make it more distinguishable.
non_process
change pagevalidator boxes background color description as per our discussion in the current boxes are similar to our input form fields let s change the background color to make it more distinguishable
0
43,285
9,415,184,758
IssuesEvent
2019-04-10 12:03:08
junhoyeo/Dimicigan-Chrome
https://api.github.com/repos/junhoyeo/Dimicigan-Chrome
closed
Generic Object Injection Sink (security/detect-object-injection)
code-style
### [Codacy](https://app.codacy.com/app/junhoyeo/dimicigan-chrome/commit?cid=339205182) detected an issue: #### Message: `Generic Object Injection Sink (security/detect-object-injection)` #### Occurred on: + **Commit**: 3b6391c93a2660fcae6b75cb3c57fcad8c46eb6c + **File**: [src/pages/Index.vue](https://github.com/junhoyeo/dimicigan-chrome/blob/3b6391c93a2660fcae6b75cb3c57fcad8c46eb6c/src/pages/Index.vue) + **LineNum**: [80](https://github.com/junhoyeo/dimicigan-chrome/blob/3b6391c93a2660fcae6b75cb3c57fcad8c46eb6c/src/pages/Index.vue#L80) + **Code**: `subject: today[i],` #### Currently on: + **Commit**: 6a2f84ade3db2d54eda829ca26cee9147d0407fb + **File**: [src/pages/Index.vue](https://github.com/junhoyeo/dimicigan-chrome/blob/6a2f84ade3db2d54eda829ca26cee9147d0407fb/src/pages/Index.vue) + **LineNum**: [80](https://github.com/junhoyeo/dimicigan-chrome/blob/6a2f84ade3db2d54eda829ca26cee9147d0407fb/src/pages/Index.vue#L80)
1.0
Generic Object Injection Sink (security/detect-object-injection) - ### [Codacy](https://app.codacy.com/app/junhoyeo/dimicigan-chrome/commit?cid=339205182) detected an issue: #### Message: `Generic Object Injection Sink (security/detect-object-injection)` #### Occurred on: + **Commit**: 3b6391c93a2660fcae6b75cb3c57fcad8c46eb6c + **File**: [src/pages/Index.vue](https://github.com/junhoyeo/dimicigan-chrome/blob/3b6391c93a2660fcae6b75cb3c57fcad8c46eb6c/src/pages/Index.vue) + **LineNum**: [80](https://github.com/junhoyeo/dimicigan-chrome/blob/3b6391c93a2660fcae6b75cb3c57fcad8c46eb6c/src/pages/Index.vue#L80) + **Code**: `subject: today[i],` #### Currently on: + **Commit**: 6a2f84ade3db2d54eda829ca26cee9147d0407fb + **File**: [src/pages/Index.vue](https://github.com/junhoyeo/dimicigan-chrome/blob/6a2f84ade3db2d54eda829ca26cee9147d0407fb/src/pages/Index.vue) + **LineNum**: [80](https://github.com/junhoyeo/dimicigan-chrome/blob/6a2f84ade3db2d54eda829ca26cee9147d0407fb/src/pages/Index.vue#L80)
non_process
generic object injection sink security detect object injection detected an issue message generic object injection sink security detect object injection occurred on commit file linenum code subject today currently on commit file linenum
0
13,976
16,748,304,175
IssuesEvent
2021-06-11 18:38:32
sysflow-telemetry/sf-docs
https://api.github.com/repos/sysflow-telemetry/sf-docs
opened
Add a command-line flag for the processor to parse configuration and policy files as a standalone tool
enhancement sf-processor
**Indicate project** Processor **Describe the feature you'd like** A command-line flag for the processor to parse configuration and policy files for syntax errors. Example: $> sfprocessor -log=info -config=pipeline.json `-test`
1.0
Add a command-line flag for the processor to parse configuration and policy files as a standalone tool - **Indicate project** Processor **Describe the feature you'd like** A command-line flag for the processor to parse configuration and policy files for syntax errors. Example: $> sfprocessor -log=info -config=pipeline.json `-test`
process
add a command line flag for the processor to parse configuration and policy files as a standalone tool indicate project processor describe the feature you d like a command line flag for the processor to parse configuration and policy files for syntax errors example sfprocessor log info config pipeline json test
1
260,529
27,784,521,620
IssuesEvent
2023-03-17 01:14:44
DavidSpek/kubeflow
https://api.github.com/repos/DavidSpek/kubeflow
opened
CVE-2023-28155 (Medium) detected in request-2.88.0.tgz, request-2.88.2.tgz
Mend: dependency security vulnerability
## CVE-2023-28155 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>request-2.88.0.tgz</b>, <b>request-2.88.2.tgz</b></p></summary> <p> <details><summary><b>request-2.88.0.tgz</b></p></summary> <p>Simplified HTTP request client.</p> <p>Library home page: <a href="https://registry.npmjs.org/request/-/request-2.88.0.tgz">https://registry.npmjs.org/request/-/request-2.88.0.tgz</a></p> <p>Path to dependency file: /components/crud-web-apps/volumes/frontend/package.json</p> <p>Path to vulnerable library: /components/crud-web-apps/volumes/frontend/node_modules/request/package.json,/components/centraldashboard/node_modules/request/package.json</p> <p> Dependency Hierarchy: - client-node-0.12.3.tgz (Root Library) - :x: **request-2.88.0.tgz** (Vulnerable Library) </details> <details><summary><b>request-2.88.2.tgz</b></p></summary> <p>Simplified HTTP request client.</p> <p>Library home page: <a href="https://registry.npmjs.org/request/-/request-2.88.2.tgz">https://registry.npmjs.org/request/-/request-2.88.2.tgz</a></p> <p>Path to dependency file: /components/crud-web-apps/common/frontend/kubeflow-common-lib/package.json</p> <p>Path to vulnerable library: /components/crud-web-apps/common/frontend/kubeflow-common-lib/node_modules/request/package.json,/components/crud-web-apps/tensorboards/frontend/node_modules/request/package.json,/components/crud-web-apps/jupyter/frontend/node_modules/request/package.json,/components/crud-web-apps/volumes/frontend/node_modules/node-gyp/node_modules/request/package.json</p> <p> Dependency Hierarchy: - client-node-0.12.2.tgz (Root Library) - :x: **request-2.88.2.tgz** (Vulnerable Library) </details> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ** UNSUPPORTED WHEN ASSIGNED ** The Request package through 2.88.1 for Node.js allows a bypass of SSRF mitigations via an attacker-controller server that does a cross-protocol redirect (HTTP to HTTPS, or HTTPS to HTTP). NOTE: This vulnerability only affects products that are no longer supported by the maintainer. <p>Publish Date: 2023-03-16 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-28155>CVE-2023-28155</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2023-28155 (Medium) detected in request-2.88.0.tgz, request-2.88.2.tgz - ## CVE-2023-28155 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>request-2.88.0.tgz</b>, <b>request-2.88.2.tgz</b></p></summary> <p> <details><summary><b>request-2.88.0.tgz</b></p></summary> <p>Simplified HTTP request client.</p> <p>Library home page: <a href="https://registry.npmjs.org/request/-/request-2.88.0.tgz">https://registry.npmjs.org/request/-/request-2.88.0.tgz</a></p> <p>Path to dependency file: /components/crud-web-apps/volumes/frontend/package.json</p> <p>Path to vulnerable library: /components/crud-web-apps/volumes/frontend/node_modules/request/package.json,/components/centraldashboard/node_modules/request/package.json</p> <p> Dependency Hierarchy: - client-node-0.12.3.tgz (Root Library) - :x: **request-2.88.0.tgz** (Vulnerable Library) </details> <details><summary><b>request-2.88.2.tgz</b></p></summary> <p>Simplified HTTP request client.</p> <p>Library home page: <a href="https://registry.npmjs.org/request/-/request-2.88.2.tgz">https://registry.npmjs.org/request/-/request-2.88.2.tgz</a></p> <p>Path to dependency file: /components/crud-web-apps/common/frontend/kubeflow-common-lib/package.json</p> <p>Path to vulnerable library: /components/crud-web-apps/common/frontend/kubeflow-common-lib/node_modules/request/package.json,/components/crud-web-apps/tensorboards/frontend/node_modules/request/package.json,/components/crud-web-apps/jupyter/frontend/node_modules/request/package.json,/components/crud-web-apps/volumes/frontend/node_modules/node-gyp/node_modules/request/package.json</p> <p> Dependency Hierarchy: - client-node-0.12.2.tgz (Root Library) - :x: **request-2.88.2.tgz** (Vulnerable Library) </details> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ** UNSUPPORTED WHEN ASSIGNED ** The Request package through 2.88.1 for Node.js allows a bypass of SSRF mitigations via an attacker-controller server that does a cross-protocol redirect (HTTP to HTTPS, or HTTPS to HTTP). NOTE: This vulnerability only affects products that are no longer supported by the maintainer. <p>Publish Date: 2023-03-16 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-28155>CVE-2023-28155</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in request tgz request tgz cve medium severity vulnerability vulnerable libraries request tgz request tgz request tgz simplified http request client library home page a href path to dependency file components crud web apps volumes frontend package json path to vulnerable library components crud web apps volumes frontend node modules request package json components centraldashboard node modules request package json dependency hierarchy client node tgz root library x request tgz vulnerable library request tgz simplified http request client library home page a href path to dependency file components crud web apps common frontend kubeflow common lib package json path to vulnerable library components crud web apps common frontend kubeflow common lib node modules request package json components crud web apps tensorboards frontend node modules request package json components crud web apps jupyter frontend node modules request package json components crud web apps volumes frontend node modules node gyp node modules request package json dependency hierarchy client node tgz root library x request tgz vulnerable library found in base branch master vulnerability details unsupported when assigned the request package through for node js allows a bypass of ssrf mitigations via an attacker controller server that does a cross protocol redirect http to https or https to http note this vulnerability only affects products that are no longer supported by the maintainer publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with mend
0
5,590
3,251,785,631
IssuesEvent
2015-10-19 11:52:14
robertocarroll/ideas
https://api.github.com/repos/robertocarroll/ideas
opened
Stories around photos
code writing
Writing around photos, Barthes, see also http://www.thehypertext.com/2015/04/11/word-camera/ 1. Found photos from parents' album 2. Process them with Clarifai API 3. Use tags as inspiration to write - short things - 4. Put the writing back with the images
1.0
Stories around photos - Writing around photos, Barthes, see also http://www.thehypertext.com/2015/04/11/word-camera/ 1. Found photos from parents' album 2. Process them with Clarifai API 3. Use tags as inspiration to write - short things - 4. Put the writing back with the images
non_process
stories around photos writing around photos barthes see also found photos from parents album process them with clarifai api use tags as inspiration to write short things put the writing back with the images
0
277,635
24,090,807,673
IssuesEvent
2022-09-19 14:37:17
eclipse-openj9/openj9
https://api.github.com/repos/eclipse-openj9/openj9
opened
jdk19 MiniMix_aot_5m_0 hang
test failure
ERROR: type should be string, got "https://openj9-jenkins.osuosl.org/job/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/8\r\nMiniMix_aot_5m_0\r\n\r\nThere is a javacore and core created after the hang, but also another core.1202972 which isn't remained. Likely caused by sending signals since it's create a minute after the renamed core file.\r\n\r\nhttps://openj9-artifactory.osuosl.org/artifactory/ci-openj9/Test/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/8/system_test_output.tar.gz\r\n\r\n```\r\nLT 06:01:05.015 - Completed 20.0%. Number of tests started=10571 (+4241)\r\nSTF 06:05:03.037 - Heartbeat: Process LT is still running\r\nSTF 06:10:03.284 - Heartbeat: Process LT is still running\r\nSTF 06:15:03.166 - Heartbeat: Process LT is still running\r\nSTF 06:20:03.478 - Heartbeat: Process LT is still running\r\nSTF 06:25:03.249 - Heartbeat: Process LT is still running\r\nSTF 06:30:03.295 - Heartbeat: Process LT is still running\r\nSTF 06:35:03.021 - Heartbeat: Process LT is still running\r\nSTF 06:40:03.236 - Heartbeat: Process LT is still running\r\nSTF 06:45:03.344 - Heartbeat: Process LT is still running\r\nSTF 06:50:03.261 - Heartbeat: Process LT is still running\r\nSTF 06:55:03.194 - Heartbeat: Process LT is still running\r\nSTF 07:00:03.453 - Heartbeat: Process LT is still running\r\nSTF 07:05:03.250 - Heartbeat: Process LT is still running\r\nSTF 07:05:04.253 - **FAILED** Process LT has timed out\r\nSTF 07:05:04.254 - Collecting dumps for: LT \r\nSTF 07:05:04.254 - Sending SIG 3 to the java process to generate a javacore\r\nSTF 07:05:04.255 - Running command: kill -3 1202972\r\nSTF 07:05:04.255 - Redirecting stderr to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_3.stderr\r\nSTF 07:05:04.255 - Redirecting stdout to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_3.stdout\r\nSTF 07:05:04.256 - Pausing for 30 seconds\r\nSTF 07:05:34.259 - Sending SIG 3 to the java process to generate a javacore\r\nSTF 07:05:34.259 - Running command: kill -3 1202972\r\nSTF 07:05:34.259 - Redirecting stderr to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_3.stderr\r\nSTF 07:05:34.259 - Redirecting stdout to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_3.stdout\r\nSTF 07:05:34.260 - Pausing for 30 seconds\r\nSTF 07:06:04.262 - Sending SIG 3 to the java process to generate a javacore\r\nSTF 07:06:04.310 - Running command: kill -3 1202972\r\nSTF 07:06:04.310 - Redirecting stderr to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_3.stderr\r\nSTF 07:06:04.310 - Redirecting stdout to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_3.stdout\r\nSTF 07:06:04.311 - Pausing for 30 seconds\r\nSTF 07:06:34.311 - Sending SIGABRT (kill -6) to the java process to generate a core\r\nSTF 07:06:34.312 - Running command: kill -6 1202972\r\nSTF 07:06:34.312 - Redirecting stderr to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_6.stderr\r\nSTF 07:06:34.312 - Redirecting stdout to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_6.stdout\r\nSTF 07:06:34.313 - Pausing for 30 seconds\r\nSTF 07:07:04.315 - Sending SIGXCPU (kill -24) to the java process to generate an OS dump\r\nSTF 07:07:04.315 - Running command: kill -24 1202972\r\nSTF 07:07:04.315 - Redirecting stderr to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_24.stderr\r\nSTF 07:07:04.315 - Redirecting stdout to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_24.stdout\r\nLT stderr JVMDUMP039I Processing dump event \"user\", detail \"\" at 2022/09/17 07:05:04 - please wait.\r\nLT stderr JVMDUMP032I JVM requested System dump using '/home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/core.20220917.070504.1202972.0001.dmp' in response to an event\r\nLT stderr JVMDUMP010I System dump written to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/core.20220917.070504.1202972.0001.dmp\r\nLT stderr JVMDUMP032I JVM requested Java dump using '/home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/javacore.20220917.070504.1202972.0002.txt' in response to an event\r\nLT stderr JVMDUMP010I Java dump written to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/javacore.20220917.070504.1202972.0002.txt\r\nLT stderr JVMDUMP013I Processed dump event \"user\", detail \"\".\r\n```"
1.0
jdk19 MiniMix_aot_5m_0 hang - https://openj9-jenkins.osuosl.org/job/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/8 MiniMix_aot_5m_0 There is a javacore and core created after the hang, but also another core.1202972 which isn't remained. Likely caused by sending signals since it's create a minute after the renamed core file. https://openj9-artifactory.osuosl.org/artifactory/ci-openj9/Test/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/8/system_test_output.tar.gz ``` LT 06:01:05.015 - Completed 20.0%. Number of tests started=10571 (+4241) STF 06:05:03.037 - Heartbeat: Process LT is still running STF 06:10:03.284 - Heartbeat: Process LT is still running STF 06:15:03.166 - Heartbeat: Process LT is still running STF 06:20:03.478 - Heartbeat: Process LT is still running STF 06:25:03.249 - Heartbeat: Process LT is still running STF 06:30:03.295 - Heartbeat: Process LT is still running STF 06:35:03.021 - Heartbeat: Process LT is still running STF 06:40:03.236 - Heartbeat: Process LT is still running STF 06:45:03.344 - Heartbeat: Process LT is still running STF 06:50:03.261 - Heartbeat: Process LT is still running STF 06:55:03.194 - Heartbeat: Process LT is still running STF 07:00:03.453 - Heartbeat: Process LT is still running STF 07:05:03.250 - Heartbeat: Process LT is still running STF 07:05:04.253 - **FAILED** Process LT has timed out STF 07:05:04.254 - Collecting dumps for: LT STF 07:05:04.254 - Sending SIG 3 to the java process to generate a javacore STF 07:05:04.255 - Running command: kill -3 1202972 STF 07:05:04.255 - Redirecting stderr to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_3.stderr STF 07:05:04.255 - Redirecting stdout to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_3.stdout STF 07:05:04.256 - Pausing for 30 seconds STF 07:05:34.259 - Sending SIG 3 to the java process to generate a javacore STF 07:05:34.259 - Running command: kill -3 1202972 STF 07:05:34.259 - Redirecting stderr to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_3.stderr STF 07:05:34.259 - Redirecting stdout to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_3.stdout STF 07:05:34.260 - Pausing for 30 seconds STF 07:06:04.262 - Sending SIG 3 to the java process to generate a javacore STF 07:06:04.310 - Running command: kill -3 1202972 STF 07:06:04.310 - Redirecting stderr to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_3.stderr STF 07:06:04.310 - Redirecting stdout to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_3.stdout STF 07:06:04.311 - Pausing for 30 seconds STF 07:06:34.311 - Sending SIGABRT (kill -6) to the java process to generate a core STF 07:06:34.312 - Running command: kill -6 1202972 STF 07:06:34.312 - Redirecting stderr to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_6.stderr STF 07:06:34.312 - Redirecting stdout to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_6.stdout STF 07:06:34.313 - Pausing for 30 seconds STF 07:07:04.315 - Sending SIGXCPU (kill -24) to the java process to generate an OS dump STF 07:07:04.315 - Running command: kill -24 1202972 STF 07:07:04.315 - Redirecting stderr to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_24.stderr STF 07:07:04.315 - Redirecting stdout to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_24.stdout LT stderr JVMDUMP039I Processing dump event "user", detail "" at 2022/09/17 07:05:04 - please wait. LT stderr JVMDUMP032I JVM requested System dump using '/home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/core.20220917.070504.1202972.0001.dmp' in response to an event LT stderr JVMDUMP010I System dump written to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/core.20220917.070504.1202972.0001.dmp LT stderr JVMDUMP032I JVM requested Java dump using '/home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/javacore.20220917.070504.1202972.0002.txt' in response to an event LT stderr JVMDUMP010I Java dump written to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/javacore.20220917.070504.1202972.0002.txt LT stderr JVMDUMP013I Processed dump event "user", detail "". ```
non_process
minimix aot hang minimix aot there is a javacore and core created after the hang but also another core which isn t remained likely caused by sending signals since it s create a minute after the renamed core file lt completed number of tests started stf heartbeat process lt is still running stf heartbeat process lt is still running stf heartbeat process lt is still running stf heartbeat process lt is still running stf heartbeat process lt is still running stf heartbeat process lt is still running stf heartbeat process lt is still running stf heartbeat process lt is still running stf heartbeat process lt is still running stf heartbeat process lt is still running stf heartbeat process lt is still running stf heartbeat process lt is still running stf heartbeat process lt is still running stf failed process lt has timed out stf collecting dumps for lt stf sending sig to the java process to generate a javacore stf running command kill stf redirecting stderr to home jenkins workspace test extended system linux nightly testlist aqa tests tkg output minimix aot mixedloadtest results lt kill stderr stf redirecting stdout to home jenkins workspace test extended system linux nightly testlist aqa tests tkg output minimix aot mixedloadtest results lt kill stdout stf pausing for seconds stf sending sig to the java process to generate a javacore stf running command kill stf redirecting stderr to home jenkins workspace test extended system linux nightly testlist aqa tests tkg output minimix aot mixedloadtest results lt kill stderr stf redirecting stdout to home jenkins workspace test extended system linux nightly testlist aqa tests tkg output minimix aot mixedloadtest results lt kill stdout stf pausing for seconds stf sending sig to the java process to generate a javacore stf running command kill stf redirecting stderr to home jenkins workspace test extended system linux nightly testlist aqa tests tkg output minimix aot mixedloadtest results lt kill stderr stf redirecting stdout to home jenkins workspace test extended system linux nightly testlist aqa tests tkg output minimix aot mixedloadtest results lt kill stdout stf pausing for seconds stf sending sigabrt kill to the java process to generate a core stf running command kill stf redirecting stderr to home jenkins workspace test extended system linux nightly testlist aqa tests tkg output minimix aot mixedloadtest results lt kill stderr stf redirecting stdout to home jenkins workspace test extended system linux nightly testlist aqa tests tkg output minimix aot mixedloadtest results lt kill stdout stf pausing for seconds stf sending sigxcpu kill to the java process to generate an os dump stf running command kill stf redirecting stderr to home jenkins workspace test extended system linux nightly testlist aqa tests tkg output minimix aot mixedloadtest results lt kill stderr stf redirecting stdout to home jenkins workspace test extended system linux nightly testlist aqa tests tkg output minimix aot mixedloadtest results lt kill stdout lt stderr processing dump event user detail at please wait lt stderr jvm requested system dump using home jenkins workspace test extended system linux nightly testlist aqa tests tkg output minimix aot mixedloadtest results core dmp in response to an event lt stderr system dump written to home jenkins workspace test extended system linux nightly testlist aqa tests tkg output minimix aot mixedloadtest results core dmp lt stderr jvm requested java dump using home jenkins workspace test extended system linux nightly testlist aqa tests tkg output minimix aot mixedloadtest results javacore txt in response to an event lt stderr java dump written to home jenkins workspace test extended system linux nightly testlist aqa tests tkg output minimix aot mixedloadtest results javacore txt lt stderr processed dump event user detail
0
21,169
28,140,404,000
IssuesEvent
2023-04-01 21:59:33
kserve/kserve
https://api.github.com/repos/kserve/kserve
closed
Parametrized container builds
kind/feature kserve/release-process
/kind feature **Describe the solution you'd like** Today, the base images are hardcoded in Dockerfiles, which makes it hard to rebuild the KServe components on custom images. Parametrizing the Dockerfile base images would enable KServe providers to swap base images to improve security or integration with the provider's platform.
1.0
Parametrized container builds - /kind feature **Describe the solution you'd like** Today, the base images are hardcoded in Dockerfiles, which makes it hard to rebuild the KServe components on custom images. Parametrizing the Dockerfile base images would enable KServe providers to swap base images to improve security or integration with the provider's platform.
process
parametrized container builds kind feature describe the solution you d like today the base images are hardcoded in dockerfiles which makes it hard to rebuild the kserve components on custom images parametrizing the dockerfile base images would enable kserve providers to swap base images to improve security or integration with the provider s platform
1
1,198
3,697,440,741
IssuesEvent
2016-02-27 17:39:39
pelias/fuzzy-tester
https://api.github.com/repos/pelias/fuzzy-tester
closed
Use Lat/Lon as one of the factors for fuzzy scoring
processed
One of the things that we care a lot about is "is this the right location". Right now, our testsuite doesn't take the location of the place into account, but instead relies on matching the labels of documents it knows to be correct. But often there are several entries for a single place, an artifact of us importing from multiple data sources and only checking for duplication within the dataset. These documents are of varying quality, but all that matters to the user is "Is the place right". In addition to these criteria, we should incorporate comparisons of distance from the "correct" document as well. One possibility of how to consider this is to calculate distance from the "correct" target document for each query and have a gradient for scoring [e.g 0-10m, 10-30m, 30-100m, 500m-2k, 2k+] where the closer it is to the target the higher the score.
1.0
Use Lat/Lon as one of the factors for fuzzy scoring - One of the things that we care a lot about is "is this the right location". Right now, our testsuite doesn't take the location of the place into account, but instead relies on matching the labels of documents it knows to be correct. But often there are several entries for a single place, an artifact of us importing from multiple data sources and only checking for duplication within the dataset. These documents are of varying quality, but all that matters to the user is "Is the place right". In addition to these criteria, we should incorporate comparisons of distance from the "correct" document as well. One possibility of how to consider this is to calculate distance from the "correct" target document for each query and have a gradient for scoring [e.g 0-10m, 10-30m, 30-100m, 500m-2k, 2k+] where the closer it is to the target the higher the score.
process
use lat lon as one of the factors for fuzzy scoring one of the things that we care a lot about is is this the right location right now our testsuite doesn t take the location of the place into account but instead relies on matching the labels of documents it knows to be correct but often there are several entries for a single place an artifact of us importing from multiple data sources and only checking for duplication within the dataset these documents are of varying quality but all that matters to the user is is the place right in addition to these criteria we should incorporate comparisons of distance from the correct document as well one possibility of how to consider this is to calculate distance from the correct target document for each query and have a gradient for scoring where the closer it is to the target the higher the score
1
12,258
14,787,266,031
IssuesEvent
2021-01-12 07:18:08
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[Mobile Apps] Studies list > Progress bar is not updated for study completion % in studies list
Bug P1 Process: Dev Process: Reopened Process: Tested dev
**Steps:** 1. Login to mobile 2. Enroll into a study 3. Complete some activities 4. Navigate to studies list 5. Observe the progress bar **Actual**: Progress bar is not updated for study completion % in studies list - iOS Progress bar is not updated for study completion % in studies list after logout and login - Android **Expected**: Progress bar is should be updated for study completion % in studies list ![iOS_progress1](https://user-images.githubusercontent.com/60386291/103406417-a8833400-4b80-11eb-9fb9-b81afb62c211.png) Completion % updating in Dashboard: ![Dashboard](https://user-images.githubusercontent.com/60386291/103406493-e84a1b80-4b80-11eb-8504-abe4c3ff6ae9.png)
3.0
[Mobile Apps] Studies list > Progress bar is not updated for study completion % in studies list - **Steps:** 1. Login to mobile 2. Enroll into a study 3. Complete some activities 4. Navigate to studies list 5. Observe the progress bar **Actual**: Progress bar is not updated for study completion % in studies list - iOS Progress bar is not updated for study completion % in studies list after logout and login - Android **Expected**: Progress bar is should be updated for study completion % in studies list ![iOS_progress1](https://user-images.githubusercontent.com/60386291/103406417-a8833400-4b80-11eb-9fb9-b81afb62c211.png) Completion % updating in Dashboard: ![Dashboard](https://user-images.githubusercontent.com/60386291/103406493-e84a1b80-4b80-11eb-8504-abe4c3ff6ae9.png)
process
studies list progress bar is not updated for study completion in studies list steps login to mobile enroll into a study complete some activities navigate to studies list observe the progress bar actual progress bar is not updated for study completion in studies list ios progress bar is not updated for study completion in studies list after logout and login android expected progress bar is should be updated for study completion in studies list completion updating in dashboard
1
164,701
6,254,181,696
IssuesEvent
2017-07-14 00:56:15
HabitRPG/habitica
https://api.github.com/repos/HabitRPG/habitica
closed
Mod Tools Desired
priority: important status: issue: suggestion-discussion
Figured we should open a ticket to start discussing some Moderator Tools that would make our lives easier. Some things that would be helpful to me: - Ability to edit/delete Challenges - Possibly, ability to edit messages instead of just deleting them (maybe to put trigger warnings at the beginning, etc). - Profanity filters. I know that there has been some debate about profanity filters, but it gets pretty tedious to filter out messages in the Tavern manually - and many of them are innocent mistakes. Even a popup with our stock message and a swapped in suggestion word (Curses! Pain in the neck! Dratitude!) would go a long way. Of course we'd occasionally get people going out of their way to circumnavigate it, but they're probably trolling anyway, so it would still minimize the hassle we have. - Similarly, we should have something in place to prevent users from choosing usernames that contain language that isn't acceptable in the public chat rooms. I just received an angry email from someone who chose an inappropriate username unwittingly and when asked to correct it, got very upset. She had a good point, which is that if it is inappropriate, she shouldn't have been allowed to choose it. (She also had some... less helpful.... suggestions about what we should go do, but I'm choosing to extract the constructive part of the feedback.) I'm sure there are plenty of other useful tools - these are just some that I've been wanting a lot recently. Other mods should definitely chime in with what they want! @Alys @deilann @DanielTheBard @veeeeeee @lefnire @SabreCat @paglias ## <bountysource-plugin> Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/5197741-mod-tools-desired?utm_campaign=plugin&utm_content=tracker%2F68393&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F68393&utm_medium=issues&utm_source=github). </bountysource-plugin>
1.0
Mod Tools Desired - Figured we should open a ticket to start discussing some Moderator Tools that would make our lives easier. Some things that would be helpful to me: - Ability to edit/delete Challenges - Possibly, ability to edit messages instead of just deleting them (maybe to put trigger warnings at the beginning, etc). - Profanity filters. I know that there has been some debate about profanity filters, but it gets pretty tedious to filter out messages in the Tavern manually - and many of them are innocent mistakes. Even a popup with our stock message and a swapped in suggestion word (Curses! Pain in the neck! Dratitude!) would go a long way. Of course we'd occasionally get people going out of their way to circumnavigate it, but they're probably trolling anyway, so it would still minimize the hassle we have. - Similarly, we should have something in place to prevent users from choosing usernames that contain language that isn't acceptable in the public chat rooms. I just received an angry email from someone who chose an inappropriate username unwittingly and when asked to correct it, got very upset. She had a good point, which is that if it is inappropriate, she shouldn't have been allowed to choose it. (She also had some... less helpful.... suggestions about what we should go do, but I'm choosing to extract the constructive part of the feedback.) I'm sure there are plenty of other useful tools - these are just some that I've been wanting a lot recently. Other mods should definitely chime in with what they want! @Alys @deilann @DanielTheBard @veeeeeee @lefnire @SabreCat @paglias ## <bountysource-plugin> Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/5197741-mod-tools-desired?utm_campaign=plugin&utm_content=tracker%2F68393&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F68393&utm_medium=issues&utm_source=github). </bountysource-plugin>
non_process
mod tools desired figured we should open a ticket to start discussing some moderator tools that would make our lives easier some things that would be helpful to me ability to edit delete challenges possibly ability to edit messages instead of just deleting them maybe to put trigger warnings at the beginning etc profanity filters i know that there has been some debate about profanity filters but it gets pretty tedious to filter out messages in the tavern manually and many of them are innocent mistakes even a popup with our stock message and a swapped in suggestion word curses pain in the neck dratitude would go a long way of course we d occasionally get people going out of their way to circumnavigate it but they re probably trolling anyway so it would still minimize the hassle we have similarly we should have something in place to prevent users from choosing usernames that contain language that isn t acceptable in the public chat rooms i just received an angry email from someone who chose an inappropriate username unwittingly and when asked to correct it got very upset she had a good point which is that if it is inappropriate she shouldn t have been allowed to choose it she also had some less helpful suggestions about what we should go do but i m choosing to extract the constructive part of the feedback i m sure there are plenty of other useful tools these are just some that i ve been wanting a lot recently other mods should definitely chime in with what they want alys deilann danielthebard veeeeeee lefnire sabrecat paglias want to back this issue we accept bounties via
0
13,461
15,946,025,585
IssuesEvent
2021-04-15 00:03:51
googleapis/release-please
https://api.github.com/repos/googleapis/release-please
closed
add tests for each language's releaser
type: process
we've split out the release logic into specific languages, e.g., ruby-yoshi, node, let's actually add some tests for each language.
1.0
add tests for each language's releaser - we've split out the release logic into specific languages, e.g., ruby-yoshi, node, let's actually add some tests for each language.
process
add tests for each language s releaser we ve split out the release logic into specific languages e g ruby yoshi node let s actually add some tests for each language
1
331,555
24,312,567,958
IssuesEvent
2022-09-30 00:56:38
roots/bud
https://api.github.com/repos/roots/bud
closed
[bug] bud-sass loader not registering, ignores files using a transpiler source
documentation
### Agreement - [X] This is not a duplicate of an existing issue - [X] I have read the [guidelines for Contributing to Roots Projects](https://github.com/roots/.github/blob/master/CONTRIBUTING.md) - [X] This is not a personal support request that should be posted on the [Roots Discourse](https://discourse.roots.io/) community ### Describe the issue I have a need to load styles/js from other directories than `./resources`. I wish to compile .scss files in a directory to their own CSS files. I have created a test directory in the root of my sage theme, containing an empty `script.js` and an empty `style.scss` I then add the test folder to the entry point: ```js app /** * Application entrypoints */ .entry({ app: ["@scripts/app", "@styles/app"], editor: ["@scripts/editor", "@styles/editor"], test: [app.path("./test/script.js"), app.path("./test/style.scss")] }) ``` After that, I add a transpiler source to the js and sass rules: ```js app.build.rules.js.setInclude([ bud => bud.path('@src'), bud => bud.path('./test'), ]) ``` This works fine, but the .scss file is ignored. So I go ahead and try add another source to the sass rule: ```js app.build.rules.sass.setInclude([ bud => bud.path('@src'), bud => bud.path('./test'), ]) ``` Which throws an error: ` Type Error: Cannot read properties of undefined (reading 'setInclude') ` In my bud config, I console log out the `app.build.rules` and see there is no sass rule: ```js { yml: Rule { _app: [Function (anonymous)], include: [ [Function (anonymous)] ], test: [Function (anonymous)], use: [ 'yml' ] }, webp: Rule { _app: [Function (anonymous)], test: [Function (anonymous)], include: [ [Function (anonymous)] ], type: 'asset/resource', generator: [Function (anonymous)] }, svg: Rule { _app: [Function (anonymous)], test: [Function (anonymous)], include: [ [Function (anonymous)] ], type: 'asset/resource', generator: [Function: bound svgGenerator] }, json: Rule { _app: [Function (anonymous)], type: 'json', include: [ [Function (anonymous)] ], test: [Function (anonymous)], parser: [Function (anonymous)] }, js: Rule { _app: [Function (anonymous)], test: [Function (anonymous)], include: [ [Function (anonymous)] ], use: [ [Item] ] }, image: Rule { _app: [Function (anonymous)], test: [Function (anonymous)], include: [ [Function (anonymous)] ], type: 'asset/resource', generator: [Function (anonymous)] }, html: Rule { _app: [Function (anonymous)], include: [ [Function (anonymous)] ], test: [Function (anonymous)], use: [ 'html' ] }, font: Rule { _app: [Function (anonymous)], type: 'asset', test: [Function (anonymous)], include: [ [Function (anonymous)] ], generator: [Function (anonymous)] }, cssModule: Rule { _app: [Function (anonymous)], test: [Function (anonymous)], include: [ [Function (anonymous)] ], use: [ 'precss', 'cssModule', 'postcss' ] }, css: Rule { _app: [Function (anonymous)], test: [Function (anonymous)], include: [ [Function (anonymous)] ], use: [ 'precss', 'css', 'postcss' ] } } ``` ### Expected Behavior .scss files in additional entry points, that have a transpiler source - to be parsed using bud-sass ### Actual Behavior This error is thrown: `Type Error: Cannot read properties of undefined (reading 'setInclude')` ### Steps To Reproduce 1. Create a test folder in the root of the teme 2. Add a blank script.js and style.scss to the folder 3. Add the entry points for the test folder into the bud config 4. Add transpiler sources for the js/sass rules ### version 6.4.4 ### What package manager are you using? npm ### version 8.15.0 ### Logs ```zsh Type Error: Cannot read properties of undefined (reading 'setInclude') at default (file:///Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/bud.config.mjs:19:24) at Configuration.run (file:///Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/node_modules/@roots/bud-framework/lib/configuration/configuration.js:30:26) at file:///Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/node_modules/@roots/bud-framework/lib/configuration/index.js:14:29 at Array.map (<anonymous>) at Module.process (file:///Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/node_modules/@roots/bud-framework/lib/configuration/index.js:13:60) at Bud.run (file:///Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/node_modules/@roots/bud-framework/lib/methods/run.js:5:29) at BuildDevelopmentCommand.runCommand (file:///Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/node_modules/@roots/bud/lib/cli/commands/build.base.js:192:24) at BuildDevelopmentCommand.execute (file:///Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/node_modules/@roots/bud/lib/cli/commands/base.js:179:24) at async BuildDevelopmentCommand.validateAndExecute (/Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/node_modules/clipanion/lib/advanced/Command.js:73:26) at async Cli.run (/Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/node_modules/clipanion/lib/advanced/Cli.js:222:24) ``` ### Configuration ```zsh // @ts-check /** * Build configuration * * @see {@link https://bud.js.org/guides/configure} * @param {import('@roots/bud').Bud} app */ export default async (app) => { console.log(app.build.rules) app.build.rules.js.setInclude([ bud => bud.path('@src'), bud => bud.path('./test'), ]) app.build.rules.scss.setInclude([ bud => bud.path('@src'), bud => bud.path('./test'), ]) app /** * Application entrypoints */ .entry({ app: ["@scripts/app", "@styles/app"], editor: ["@scripts/editor", "@styles/editor"], test: [app.path("./test/script.js"), app.path("./test/style.scss")] }) /** * Directory contents to be included in the compilation */ .assets(["images"]) /** * Matched files trigger a page reload when modified */ .watch(["resources/views/**/*", "app/**/*"]) /** * Proxy origin (`WP_HOME`) */ .proxy("http://example.test") /** * Development origin */ .serve("http://0.0.0.0:3000") /** * URI of the `public` directory */ .setPublicPath("/app/themes/sage/public/") /** * Generate WordPress `theme.json` * * @note This overwrites `theme.json` on every build. */ .wpjson .settings({ color: { custom: false, customGradient: false, defaultPalette: false, defaultGradients: false, }, custom: { spacing: {}, typography: { 'font-size': {}, 'line-height': {}, }, }, spacing: { padding: true, units: ['px', '%', 'em', 'rem', 'vw', 'vh'], }, typography: { customFontSize: false, }, }) .useTailwindColors() .useTailwindFontFamily() .useTailwindFontSize() .enable() }; ``` ### Relevant .budfiles _No response_
1.0
[bug] bud-sass loader not registering, ignores files using a transpiler source - ### Agreement - [X] This is not a duplicate of an existing issue - [X] I have read the [guidelines for Contributing to Roots Projects](https://github.com/roots/.github/blob/master/CONTRIBUTING.md) - [X] This is not a personal support request that should be posted on the [Roots Discourse](https://discourse.roots.io/) community ### Describe the issue I have a need to load styles/js from other directories than `./resources`. I wish to compile .scss files in a directory to their own CSS files. I have created a test directory in the root of my sage theme, containing an empty `script.js` and an empty `style.scss` I then add the test folder to the entry point: ```js app /** * Application entrypoints */ .entry({ app: ["@scripts/app", "@styles/app"], editor: ["@scripts/editor", "@styles/editor"], test: [app.path("./test/script.js"), app.path("./test/style.scss")] }) ``` After that, I add a transpiler source to the js and sass rules: ```js app.build.rules.js.setInclude([ bud => bud.path('@src'), bud => bud.path('./test'), ]) ``` This works fine, but the .scss file is ignored. So I go ahead and try add another source to the sass rule: ```js app.build.rules.sass.setInclude([ bud => bud.path('@src'), bud => bud.path('./test'), ]) ``` Which throws an error: ` Type Error: Cannot read properties of undefined (reading 'setInclude') ` In my bud config, I console log out the `app.build.rules` and see there is no sass rule: ```js { yml: Rule { _app: [Function (anonymous)], include: [ [Function (anonymous)] ], test: [Function (anonymous)], use: [ 'yml' ] }, webp: Rule { _app: [Function (anonymous)], test: [Function (anonymous)], include: [ [Function (anonymous)] ], type: 'asset/resource', generator: [Function (anonymous)] }, svg: Rule { _app: [Function (anonymous)], test: [Function (anonymous)], include: [ [Function (anonymous)] ], type: 'asset/resource', generator: [Function: bound svgGenerator] }, json: Rule { _app: [Function (anonymous)], type: 'json', include: [ [Function (anonymous)] ], test: [Function (anonymous)], parser: [Function (anonymous)] }, js: Rule { _app: [Function (anonymous)], test: [Function (anonymous)], include: [ [Function (anonymous)] ], use: [ [Item] ] }, image: Rule { _app: [Function (anonymous)], test: [Function (anonymous)], include: [ [Function (anonymous)] ], type: 'asset/resource', generator: [Function (anonymous)] }, html: Rule { _app: [Function (anonymous)], include: [ [Function (anonymous)] ], test: [Function (anonymous)], use: [ 'html' ] }, font: Rule { _app: [Function (anonymous)], type: 'asset', test: [Function (anonymous)], include: [ [Function (anonymous)] ], generator: [Function (anonymous)] }, cssModule: Rule { _app: [Function (anonymous)], test: [Function (anonymous)], include: [ [Function (anonymous)] ], use: [ 'precss', 'cssModule', 'postcss' ] }, css: Rule { _app: [Function (anonymous)], test: [Function (anonymous)], include: [ [Function (anonymous)] ], use: [ 'precss', 'css', 'postcss' ] } } ``` ### Expected Behavior .scss files in additional entry points, that have a transpiler source - to be parsed using bud-sass ### Actual Behavior This error is thrown: `Type Error: Cannot read properties of undefined (reading 'setInclude')` ### Steps To Reproduce 1. Create a test folder in the root of the teme 2. Add a blank script.js and style.scss to the folder 3. Add the entry points for the test folder into the bud config 4. Add transpiler sources for the js/sass rules ### version 6.4.4 ### What package manager are you using? npm ### version 8.15.0 ### Logs ```zsh Type Error: Cannot read properties of undefined (reading 'setInclude') at default (file:///Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/bud.config.mjs:19:24) at Configuration.run (file:///Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/node_modules/@roots/bud-framework/lib/configuration/configuration.js:30:26) at file:///Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/node_modules/@roots/bud-framework/lib/configuration/index.js:14:29 at Array.map (<anonymous>) at Module.process (file:///Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/node_modules/@roots/bud-framework/lib/configuration/index.js:13:60) at Bud.run (file:///Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/node_modules/@roots/bud-framework/lib/methods/run.js:5:29) at BuildDevelopmentCommand.runCommand (file:///Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/node_modules/@roots/bud/lib/cli/commands/build.base.js:192:24) at BuildDevelopmentCommand.execute (file:///Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/node_modules/@roots/bud/lib/cli/commands/base.js:179:24) at async BuildDevelopmentCommand.validateAndExecute (/Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/node_modules/clipanion/lib/advanced/Command.js:73:26) at async Cli.run (/Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/node_modules/clipanion/lib/advanced/Cli.js:222:24) ``` ### Configuration ```zsh // @ts-check /** * Build configuration * * @see {@link https://bud.js.org/guides/configure} * @param {import('@roots/bud').Bud} app */ export default async (app) => { console.log(app.build.rules) app.build.rules.js.setInclude([ bud => bud.path('@src'), bud => bud.path('./test'), ]) app.build.rules.scss.setInclude([ bud => bud.path('@src'), bud => bud.path('./test'), ]) app /** * Application entrypoints */ .entry({ app: ["@scripts/app", "@styles/app"], editor: ["@scripts/editor", "@styles/editor"], test: [app.path("./test/script.js"), app.path("./test/style.scss")] }) /** * Directory contents to be included in the compilation */ .assets(["images"]) /** * Matched files trigger a page reload when modified */ .watch(["resources/views/**/*", "app/**/*"]) /** * Proxy origin (`WP_HOME`) */ .proxy("http://example.test") /** * Development origin */ .serve("http://0.0.0.0:3000") /** * URI of the `public` directory */ .setPublicPath("/app/themes/sage/public/") /** * Generate WordPress `theme.json` * * @note This overwrites `theme.json` on every build. */ .wpjson .settings({ color: { custom: false, customGradient: false, defaultPalette: false, defaultGradients: false, }, custom: { spacing: {}, typography: { 'font-size': {}, 'line-height': {}, }, }, spacing: { padding: true, units: ['px', '%', 'em', 'rem', 'vw', 'vh'], }, typography: { customFontSize: false, }, }) .useTailwindColors() .useTailwindFontFamily() .useTailwindFontSize() .enable() }; ``` ### Relevant .budfiles _No response_
non_process
bud sass loader not registering ignores files using a transpiler source agreement this is not a duplicate of an existing issue i have read the this is not a personal support request that should be posted on the community describe the issue i have a need to load styles js from other directories than resources i wish to compile scss files in a directory to their own css files i have created a test directory in the root of my sage theme containing an empty script js and an empty style scss i then add the test folder to the entry point js app application entrypoints entry app editor test after that i add a transpiler source to the js and sass rules js app build rules js setinclude bud bud path src bud bud path test this works fine but the scss file is ignored so i go ahead and try add another source to the sass rule js app build rules sass setinclude bud bud path src bud bud path test which throws an error type error cannot read properties of undefined reading setinclude in my bud config i console log out the app build rules and see there is no sass rule js yml rule app include test use webp rule app test include type asset resource generator svg rule app test include type asset resource generator json rule app type json include test parser js rule app test include use image rule app test include type asset resource generator html rule app include test use font rule app type asset test include generator cssmodule rule app test include use css rule app test include use expected behavior scss files in additional entry points that have a transpiler source to be parsed using bud sass actual behavior this error is thrown type error cannot read properties of undefined reading setinclude steps to reproduce create a test folder in the root of the teme add a blank script js and style scss to the folder add the entry points for the test folder into the bud config add transpiler sources for the js sass rules version what package manager are you using npm version logs zsh type error cannot read properties of undefined reading setinclude at default file users marcbroad desktop www sallys site web app themes sage bud config mjs at configuration run file users marcbroad desktop www sallys site web app themes sage node modules roots bud framework lib configuration configuration js at file users marcbroad desktop www sallys site web app themes sage node modules roots bud framework lib configuration index js at array map at module process file users marcbroad desktop www sallys site web app themes sage node modules roots bud framework lib configuration index js at bud run file users marcbroad desktop www sallys site web app themes sage node modules roots bud framework lib methods run js at builddevelopmentcommand runcommand file users marcbroad desktop www sallys site web app themes sage node modules roots bud lib cli commands build base js at builddevelopmentcommand execute file users marcbroad desktop www sallys site web app themes sage node modules roots bud lib cli commands base js at async builddevelopmentcommand validateandexecute users marcbroad desktop www sallys site web app themes sage node modules clipanion lib advanced command js at async cli run users marcbroad desktop www sallys site web app themes sage node modules clipanion lib advanced cli js configuration zsh ts check build configuration see link param import roots bud bud app export default async app console log app build rules app build rules js setinclude bud bud path src bud bud path test app build rules scss setinclude bud bud path src bud bud path test app application entrypoints entry app editor test directory contents to be included in the compilation assets matched files trigger a page reload when modified watch proxy origin wp home proxy development origin serve uri of the public directory setpublicpath app themes sage public generate wordpress theme json note this overwrites theme json on every build wpjson settings color custom false customgradient false defaultpalette false defaultgradients false custom spacing typography font size line height spacing padding true units typography customfontsize false usetailwindcolors usetailwindfontfamily usetailwindfontsize enable relevant budfiles no response
0
11,419
14,246,202,566
IssuesEvent
2020-11-19 09:45:10
freedomofpress/securedrop
https://api.github.com/repos/freedomofpress/securedrop
closed
i18n_tool.py list-translators is broken by multiple syncs in a release
goals: speed up release process
## Description The `list-translators` function has a brittle heuristic for determining when to start gathering translator contributions; it looks backward through the git history for the most recent `l10n: sync` message. If we have to sync more than once for a release, to incorporate source string feedback or whatever, then it will miss contributions since the last release and before that sync.
1.0
i18n_tool.py list-translators is broken by multiple syncs in a release - ## Description The `list-translators` function has a brittle heuristic for determining when to start gathering translator contributions; it looks backward through the git history for the most recent `l10n: sync` message. If we have to sync more than once for a release, to incorporate source string feedback or whatever, then it will miss contributions since the last release and before that sync.
process
tool py list translators is broken by multiple syncs in a release description the list translators function has a brittle heuristic for determining when to start gathering translator contributions it looks backward through the git history for the most recent sync message if we have to sync more than once for a release to incorporate source string feedback or whatever then it will miss contributions since the last release and before that sync
1
8,292
11,458,273,057
IssuesEvent
2020-02-07 02:44:53
googleapis/google-cloud-python
https://api.github.com/repos/googleapis/google-cloud-python
closed
Generated noxfiles missing 'cover' session
api: automl api: bigquerydatatransfer api: cloudasset api: cloudiot api: cloudkms api: cloudtasks api: container api: dataproc api: oslogin api: texttospeech api: videointelligence testing type: process
They are also missing the `lint` session. While the argument might be made, "they are autogen-only", the fact is that a) the `gapic-generator`, `synthtool`, etc. are software, and have bugs too; b) the local `synth.py` is perfectly capable of injecting its *own* bugs. `videointelligence` is notable in this list because it is at "beta" support level, so we should expect higher code quality for it.
1.0
Generated noxfiles missing 'cover' session - They are also missing the `lint` session. While the argument might be made, "they are autogen-only", the fact is that a) the `gapic-generator`, `synthtool`, etc. are software, and have bugs too; b) the local `synth.py` is perfectly capable of injecting its *own* bugs. `videointelligence` is notable in this list because it is at "beta" support level, so we should expect higher code quality for it.
process
generated noxfiles missing cover session they are also missing the lint session while the argument might be made they are autogen only the fact is that a the gapic generator synthtool etc are software and have bugs too b the local synth py is perfectly capable of injecting its own bugs videointelligence is notable in this list because it is at beta support level so we should expect higher code quality for it
1
659,223
21,919,625,872
IssuesEvent
2022-05-22 11:21:59
kubernetes/ingress-nginx
https://api.github.com/repos/kubernetes/ingress-nginx
closed
Allow ingress controller to set default annotations for ingress resources
kind/feature lifecycle/rotten needs-triage needs-priority
When configuring an ingress controller we can set some configuration via the config map. In Helm we can use the `controlller.config` to pass the configuration to avoid having to set those setting as annotations in the ingress resource. It would be nice to also have a way to set some default annotations that all ingress resources that are using that particular `ingressClass` will inherit. For example `cert-manager.io/cluster-issuer: somevalue` could be a useful annotation to add to the ingress controller so that ingress resources don't require the annotation to be set. Is this behaviour already available somehow? Please let me know if I need to clarify the issue a bit more. THanks.
1.0
Allow ingress controller to set default annotations for ingress resources - When configuring an ingress controller we can set some configuration via the config map. In Helm we can use the `controlller.config` to pass the configuration to avoid having to set those setting as annotations in the ingress resource. It would be nice to also have a way to set some default annotations that all ingress resources that are using that particular `ingressClass` will inherit. For example `cert-manager.io/cluster-issuer: somevalue` could be a useful annotation to add to the ingress controller so that ingress resources don't require the annotation to be set. Is this behaviour already available somehow? Please let me know if I need to clarify the issue a bit more. THanks.
non_process
allow ingress controller to set default annotations for ingress resources when configuring an ingress controller we can set some configuration via the config map in helm we can use the controlller config to pass the configuration to avoid having to set those setting as annotations in the ingress resource it would be nice to also have a way to set some default annotations that all ingress resources that are using that particular ingressclass will inherit for example cert manager io cluster issuer somevalue could be a useful annotation to add to the ingress controller so that ingress resources don t require the annotation to be set is this behaviour already available somehow please let me know if i need to clarify the issue a bit more thanks
0
18,926
24,880,715,634
IssuesEvent
2022-10-28 00:37:40
apache/arrow-rs
https://api.github.com/repos/apache/arrow-rs
closed
Make take kernel not take values of childs when taking a null
arrow development-process
*Note*: migrated from original JIRA: https://issues.apache.org/jira/browse/ARROW-10594 Currently, take just takes all values from the childs, irrespectively of whether we took a null or not.
1.0
Make take kernel not take values of childs when taking a null - *Note*: migrated from original JIRA: https://issues.apache.org/jira/browse/ARROW-10594 Currently, take just takes all values from the childs, irrespectively of whether we took a null or not.
process
make take kernel not take values of childs when taking a null note migrated from original jira currently take just takes all values from the childs irrespectively of whether we took a null or not
1
418,056
28,113,374,333
IssuesEvent
2023-03-31 08:56:10
tangphi/ped
https://api.github.com/repos/tangphi/ped
opened
Not enough visuals in UG
severity.VeryLow type.DocumentationBug
![Screenshot 2023-03-31 at 4.54.27 pm.png](https://raw.githubusercontent.com/tangphi/ped/main/files/a0353a31-d751-455b-a44f-bc957ab38e8a.png) Since the graph feature outputs a graph, it would be nice to have an example output based on the given example input. <!--session: 1680252431019-08cca115-f668-44e8-9c9b-d3d5514bd741--> <!--Version: Web v3.4.7-->
1.0
Not enough visuals in UG - ![Screenshot 2023-03-31 at 4.54.27 pm.png](https://raw.githubusercontent.com/tangphi/ped/main/files/a0353a31-d751-455b-a44f-bc957ab38e8a.png) Since the graph feature outputs a graph, it would be nice to have an example output based on the given example input. <!--session: 1680252431019-08cca115-f668-44e8-9c9b-d3d5514bd741--> <!--Version: Web v3.4.7-->
non_process
not enough visuals in ug since the graph feature outputs a graph it would be nice to have an example output based on the given example input
0
22,391
31,142,286,668
IssuesEvent
2023-08-16 01:44:19
cypress-io/cypress
https://api.github.com/repos/cypress-io/cypress
closed
Flaky test: `cy.task('__internal_scaffoldProject')` failed with the following error: > EPERM: operation not permitted, stat 'C:\Users\circleci\AppData\Local\Temp\cy-projects\cypress-in-cypress'
OS: windows stage: backlog process: flaky test topic: flake ❄️ topic: scaffoldProject stale
### Link to dashboard or CircleCI failure https://app.circleci.com/pipelines/github/cypress-io/cypress/41757/workflows/1a0d6f2e-ac67-4ac6-ab24-2c00e7149ea4/jobs/1730826/tests#failed-test-0 ### Link to failing test in GitHub N/A ### Analysis <img width="1102" alt="Screen Shot 2022-08-11 at 7 55 03 PM" src="https://user-images.githubusercontent.com/26726429/184276453-d4589029-aed7-48ff-8953-7d796a8bd762.png"> ### Cypress Version 10.4.0 ### Other _No response_
1.0
Flaky test: `cy.task('__internal_scaffoldProject')` failed with the following error: > EPERM: operation not permitted, stat 'C:\Users\circleci\AppData\Local\Temp\cy-projects\cypress-in-cypress' - ### Link to dashboard or CircleCI failure https://app.circleci.com/pipelines/github/cypress-io/cypress/41757/workflows/1a0d6f2e-ac67-4ac6-ab24-2c00e7149ea4/jobs/1730826/tests#failed-test-0 ### Link to failing test in GitHub N/A ### Analysis <img width="1102" alt="Screen Shot 2022-08-11 at 7 55 03 PM" src="https://user-images.githubusercontent.com/26726429/184276453-d4589029-aed7-48ff-8953-7d796a8bd762.png"> ### Cypress Version 10.4.0 ### Other _No response_
process
flaky test cy task internal scaffoldproject failed with the following error eperm operation not permitted stat c users circleci appdata local temp cy projects cypress in cypress link to dashboard or circleci failure link to failing test in github n a analysis img width alt screen shot at pm src cypress version other no response
1
70,187
13,436,035,455
IssuesEvent
2020-09-07 13:48:21
easably/website
https://api.github.com/repos/easably/website
opened
Design two pop-up window for standard Promo Code.
design promo codes
Description: Develop two pop-up windows for standard Promo Code. The first layout is a pop-up window after choosing a free promo code (Apply Now). The second layout is the page that the customer will see after confirming the application of the promo code (Congratulation+ Your Promo code has been successfully applied). An example for design can be taken from the page: https://www.figma.com/file/oRhHhpPn92aPwa8kSFxsJ6/Untitled?node-id=0%3A1
1.0
Design two pop-up window for standard Promo Code. - Description: Develop two pop-up windows for standard Promo Code. The first layout is a pop-up window after choosing a free promo code (Apply Now). The second layout is the page that the customer will see after confirming the application of the promo code (Congratulation+ Your Promo code has been successfully applied). An example for design can be taken from the page: https://www.figma.com/file/oRhHhpPn92aPwa8kSFxsJ6/Untitled?node-id=0%3A1
non_process
design two pop up window for standard promo code description develop two pop up windows for standard promo code the first layout is a pop up window after choosing a free promo code apply now the second layout is the page that the customer will see after confirming the application of the promo code congratulation your promo code has been successfully applied an example for design can be taken from the page
0
181,961
14,894,871,680
IssuesEvent
2021-01-21 08:16:28
equinor/flownet
https://api.github.com/repos/equinor/flownet
opened
Write documentation on relperm modelling
documentation
The relative permeability modelling has in the meanwhile become quite extensive. We should write documentation on it.
1.0
Write documentation on relperm modelling - The relative permeability modelling has in the meanwhile become quite extensive. We should write documentation on it.
non_process
write documentation on relperm modelling the relative permeability modelling has in the meanwhile become quite extensive we should write documentation on it
0
18,531
24,552,697,614
IssuesEvent
2022-10-12 13:44:55
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[Mobile apps] Study activities are not getting loaded for enrolled studies
Bug Blocker P0 iOS Android Process: Fixed Process: Tested dev
Study activities are not getting loaded for enrolled studies - Getting continuous loading ![image](https://user-images.githubusercontent.com/71445210/179495423-1a1f2f78-fc70-49e8-8552-a8c084a7a99d.png)
2.0
[Mobile apps] Study activities are not getting loaded for enrolled studies - Study activities are not getting loaded for enrolled studies - Getting continuous loading ![image](https://user-images.githubusercontent.com/71445210/179495423-1a1f2f78-fc70-49e8-8552-a8c084a7a99d.png)
process
study activities are not getting loaded for enrolled studies study activities are not getting loaded for enrolled studies getting continuous loading
1
21,741
30,257,820,351
IssuesEvent
2023-07-07 05:20:34
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
Grab docker label using docker/resourcedetection
enhancement Stale processor/resourcedetection closed as inactive
### Component(s) processor/resourcedetection ### Is your feature request related to a problem? Please describe. I'm using nomad to run otel, I got difficulty when trying to get the nomad label and send it to loki. The possible way to grab nomad label is by using docker/resourcedetection. but currently, docker/resourcedetection only supports [host.name and os.type](https://github.com/open-telemetry/opentelemetry-collector-contrib/blob/main/processor/resourcedetectionprocessor/README.md#system-metadata) Here are some label from containers that might we can use to create our own label ``` "Labels": { "com.hashicorp.nomad.alloc_id": "40c4c3b8-1d1f-8f58-94ca-4efda336d62e", "com.hashicorp.nomad.job_id": "fabio", "com.hashicorp.nomad.job_name": "fabio", "com.hashicorp.nomad.namespace": "default", "com.hashicorp.nomad.node_id": "1610be6d-eea9-c4da-8d88-b945fe6e3306", "com.hashicorp.nomad.node_name": "t3a-small-ip-10-11-82-122-aws-client", "com.hashicorp.nomad.task_group_name": "fabio", "com.hashicorp.nomad.task_name": "fabio", "org.opencontainers.image.created": "2022-09-10T19:11:50Z", "org.opencontainers.image.revision": "ebf15c22df1a4b4b367a9043865d3fe51d5db013", "org.opencontainers.image.title": "fabio", "org.opencontainers.image.version": "1.6.2" } ``` ### Describe the solution you'd like Add support to use docker label in docker/resourcedetection like [vector (see vector configuration file)](https://atodorov.me/2021/07/09/logging-on-nomad-and-log-aggregation-with-loki/) ``` resourcedetection/docker: detectors: [env, docker] timeout: 2s override: false attributes: - host.name - os.type - label.com.hashicorp.nomad.job_name - label.com.hashicorp.nomad.task_name ``` ### Describe alternatives you've considered - ### Additional context -
1.0
Grab docker label using docker/resourcedetection - ### Component(s) processor/resourcedetection ### Is your feature request related to a problem? Please describe. I'm using nomad to run otel, I got difficulty when trying to get the nomad label and send it to loki. The possible way to grab nomad label is by using docker/resourcedetection. but currently, docker/resourcedetection only supports [host.name and os.type](https://github.com/open-telemetry/opentelemetry-collector-contrib/blob/main/processor/resourcedetectionprocessor/README.md#system-metadata) Here are some label from containers that might we can use to create our own label ``` "Labels": { "com.hashicorp.nomad.alloc_id": "40c4c3b8-1d1f-8f58-94ca-4efda336d62e", "com.hashicorp.nomad.job_id": "fabio", "com.hashicorp.nomad.job_name": "fabio", "com.hashicorp.nomad.namespace": "default", "com.hashicorp.nomad.node_id": "1610be6d-eea9-c4da-8d88-b945fe6e3306", "com.hashicorp.nomad.node_name": "t3a-small-ip-10-11-82-122-aws-client", "com.hashicorp.nomad.task_group_name": "fabio", "com.hashicorp.nomad.task_name": "fabio", "org.opencontainers.image.created": "2022-09-10T19:11:50Z", "org.opencontainers.image.revision": "ebf15c22df1a4b4b367a9043865d3fe51d5db013", "org.opencontainers.image.title": "fabio", "org.opencontainers.image.version": "1.6.2" } ``` ### Describe the solution you'd like Add support to use docker label in docker/resourcedetection like [vector (see vector configuration file)](https://atodorov.me/2021/07/09/logging-on-nomad-and-log-aggregation-with-loki/) ``` resourcedetection/docker: detectors: [env, docker] timeout: 2s override: false attributes: - host.name - os.type - label.com.hashicorp.nomad.job_name - label.com.hashicorp.nomad.task_name ``` ### Describe alternatives you've considered - ### Additional context -
process
grab docker label using docker resourcedetection component s processor resourcedetection is your feature request related to a problem please describe i m using nomad to run otel i got difficulty when trying to get the nomad label and send it to loki the possible way to grab nomad label is by using docker resourcedetection but currently docker resourcedetection only supports here are some label from containers that might we can use to create our own label labels com hashicorp nomad alloc id com hashicorp nomad job id fabio com hashicorp nomad job name fabio com hashicorp nomad namespace default com hashicorp nomad node id com hashicorp nomad node name small ip aws client com hashicorp nomad task group name fabio com hashicorp nomad task name fabio org opencontainers image created org opencontainers image revision org opencontainers image title fabio org opencontainers image version describe the solution you d like add support to use docker label in docker resourcedetection like resourcedetection docker detectors timeout override false attributes host name os type label com hashicorp nomad job name label com hashicorp nomad task name describe alternatives you ve considered additional context
1
233,137
25,738,940,407
IssuesEvent
2022-12-08 03:59:30
CDCgov/prime-reportstream
https://api.github.com/repos/CDCgov/prime-reportstream
closed
React App idle timer logout
security experience
The react app is a full SPA and pages load without making https requests to the server. While the REST API can detect "idle" time by delays between requests and/or can be controlled by Okta configuration(?), the user experience isn't the greatest. At some point the requests to get data would just start failing. A better approach is to use a idle detect hook that warns before idle timeout. Then after the logged out, a notice can be displayed explaining what happened. This might be a security requirement? We should verify.
True
React App idle timer logout - The react app is a full SPA and pages load without making https requests to the server. While the REST API can detect "idle" time by delays between requests and/or can be controlled by Okta configuration(?), the user experience isn't the greatest. At some point the requests to get data would just start failing. A better approach is to use a idle detect hook that warns before idle timeout. Then after the logged out, a notice can be displayed explaining what happened. This might be a security requirement? We should verify.
non_process
react app idle timer logout the react app is a full spa and pages load without making https requests to the server while the rest api can detect idle time by delays between requests and or can be controlled by okta configuration the user experience isn t the greatest at some point the requests to get data would just start failing a better approach is to use a idle detect hook that warns before idle timeout then after the logged out a notice can be displayed explaining what happened this might be a security requirement we should verify
0
7,993
2,611,071,447
IssuesEvent
2015-02-27 00:33:21
alistairreilly/andors-trail
https://api.github.com/repos/alistairreilly/andors-trail
opened
combat doesn't end after monster dies
auto-migrated Type-Defect
``` Before posting, please read the following guidelines for posts in the issue tracker: http://code.google.com/p/andors-trail/wiki/Forums_vs_issuetracker What steps will reproduce the problem? 1.I have gutherbeards dagger equipped 2.enemy loses last hp points from bleeding curse 3.I have to complete another round of attack against a dead enemy before the fight is over What is the expected output? What do you see instead? I expect the fight to end once the enemy is dead What version of the product are you using? On what device? V 0.6.12 on samsung galaxy s2 sprint Please provide any additional information below. ``` Original issue reported on code.google.com by `blue...@gmail.com` on 26 Jun 2013 at 4:38
1.0
combat doesn't end after monster dies - ``` Before posting, please read the following guidelines for posts in the issue tracker: http://code.google.com/p/andors-trail/wiki/Forums_vs_issuetracker What steps will reproduce the problem? 1.I have gutherbeards dagger equipped 2.enemy loses last hp points from bleeding curse 3.I have to complete another round of attack against a dead enemy before the fight is over What is the expected output? What do you see instead? I expect the fight to end once the enemy is dead What version of the product are you using? On what device? V 0.6.12 on samsung galaxy s2 sprint Please provide any additional information below. ``` Original issue reported on code.google.com by `blue...@gmail.com` on 26 Jun 2013 at 4:38
non_process
combat doesn t end after monster dies before posting please read the following guidelines for posts in the issue tracker what steps will reproduce the problem i have gutherbeards dagger equipped enemy loses last hp points from bleeding curse i have to complete another round of attack against a dead enemy before the fight is over what is the expected output what do you see instead i expect the fight to end once the enemy is dead what version of the product are you using on what device v on samsung galaxy sprint please provide any additional information below original issue reported on code google com by blue gmail com on jun at
0
18,007
24,024,214,291
IssuesEvent
2022-09-15 10:06:01
anitsh/til
https://api.github.com/repos/anitsh/til
opened
Guiding principle: cross-pollination over imposed standards
principle practice blog protocol process
Standards are useful to simplify learning and address variation of performance. Standards are useful to avoid everyone having to learn a new way of doing things every time they interact with a new team. Non-standard team interaction protocols ![image](https://user-images.githubusercontent.com/414141/190375304-7a312c3d-eb8f-4a91-98da-c671b7d700e8.png) Standards are useful to address variation of performance, that is, if there’s a better way of doing something, a standard can be used to spread it across teams. Re-inventing the wheel rather than spreading it via a standard The problem with imposed standards is context There are problems with imposing a centralised standard. There’s a cost to having to learn different approaches every time you switch context BUT it’s also unlikely that one approach is optimal for every context AND we want to allow for experimentation to discover even better approaches. Make defaults easy but leave the option open for alternatives Cross-pollination encourages de facto standardisation while allowing for flexibility Ensure everyone is aware of defaults but also has the autonomy to choose an alternate approach as appropriate. Make defaults easy to do (aka [Golden Path](https://engineering.atspotify.com/2020/08/how-we-use-golden-paths-to-solve-fragmentation-in-our-software-ecosystem/)) but leave the option open to choose an alternate approach as appropriate. This allows for context-specific adjustments AND experimentation to find even better approaches. ![image](https://user-images.githubusercontent.com/414141/190376177-33c9c0d1-ecdb-41c5-ac10-89bb8a24c221.png) Cross-boundary protocols generally warrants stronger guidance, even imposition Communication and interaction protocols across boundaries are where standardisation is important even if it requires more imposition. ![image](https://user-images.githubusercontent.com/414141/190376585-68500389-5ae2-40b5-900e-ee47386acf7d.png) # Resource - https://jchyip.medium.com/guiding-principle-cross-pollination-over-imposed-standards-a2375d0e8de6
1.0
Guiding principle: cross-pollination over imposed standards - Standards are useful to simplify learning and address variation of performance. Standards are useful to avoid everyone having to learn a new way of doing things every time they interact with a new team. Non-standard team interaction protocols ![image](https://user-images.githubusercontent.com/414141/190375304-7a312c3d-eb8f-4a91-98da-c671b7d700e8.png) Standards are useful to address variation of performance, that is, if there’s a better way of doing something, a standard can be used to spread it across teams. Re-inventing the wheel rather than spreading it via a standard The problem with imposed standards is context There are problems with imposing a centralised standard. There’s a cost to having to learn different approaches every time you switch context BUT it’s also unlikely that one approach is optimal for every context AND we want to allow for experimentation to discover even better approaches. Make defaults easy but leave the option open for alternatives Cross-pollination encourages de facto standardisation while allowing for flexibility Ensure everyone is aware of defaults but also has the autonomy to choose an alternate approach as appropriate. Make defaults easy to do (aka [Golden Path](https://engineering.atspotify.com/2020/08/how-we-use-golden-paths-to-solve-fragmentation-in-our-software-ecosystem/)) but leave the option open to choose an alternate approach as appropriate. This allows for context-specific adjustments AND experimentation to find even better approaches. ![image](https://user-images.githubusercontent.com/414141/190376177-33c9c0d1-ecdb-41c5-ac10-89bb8a24c221.png) Cross-boundary protocols generally warrants stronger guidance, even imposition Communication and interaction protocols across boundaries are where standardisation is important even if it requires more imposition. ![image](https://user-images.githubusercontent.com/414141/190376585-68500389-5ae2-40b5-900e-ee47386acf7d.png) # Resource - https://jchyip.medium.com/guiding-principle-cross-pollination-over-imposed-standards-a2375d0e8de6
process
guiding principle cross pollination over imposed standards standards are useful to simplify learning and address variation of performance standards are useful to avoid everyone having to learn a new way of doing things every time they interact with a new team non standard team interaction protocols standards are useful to address variation of performance that is if there’s a better way of doing something a standard can be used to spread it across teams re inventing the wheel rather than spreading it via a standard the problem with imposed standards is context there are problems with imposing a centralised standard there’s a cost to having to learn different approaches every time you switch context but it’s also unlikely that one approach is optimal for every context and we want to allow for experimentation to discover even better approaches make defaults easy but leave the option open for alternatives cross pollination encourages de facto standardisation while allowing for flexibility ensure everyone is aware of defaults but also has the autonomy to choose an alternate approach as appropriate make defaults easy to do aka but leave the option open to choose an alternate approach as appropriate this allows for context specific adjustments and experimentation to find even better approaches cross boundary protocols generally warrants stronger guidance even imposition communication and interaction protocols across boundaries are where standardisation is important even if it requires more imposition resource
1
52,193
7,752,279,930
IssuesEvent
2018-05-30 19:46:41
amberframework/granite
https://api.github.com/repos/amberframework/granite
closed
Validation helpers
kind:documentation kind:enhancement pr:needs-review
As of now the validators usage is not documented in the readme. We should do that.
1.0
Validation helpers - As of now the validators usage is not documented in the readme. We should do that.
non_process
validation helpers as of now the validators usage is not documented in the readme we should do that
0
16,663
21,731,104,807
IssuesEvent
2022-05-11 12:06:09
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Crash when running check validity using QGIS method
Processing Bug Crash/Data Corruption
### What is the bug or the crash? When running "Check Validity" on a particular geopackage using the QGIS method, QGIS crashes. When using the GEOS method, it reports no errors. I've also tested on a clean user profile, and the crash still occurs. I've isolated the problem to this single multi-polygon feature in a geopackage: [checkvalidity_crash.zip](https://github.com/qgis/QGIS/files/8663128/checkvalidity_crash.zip) ## Report Details **Python Stack Trace** ``` Windows fatal exception: access violation Current thread 0x00024620 (most recent call first): File "C:\OSGeo4W/apps/qgis/./python/plugins\processing\algs\qgis\CheckValidity.py", line 160 in doCheck errors = list(geom.validateGeometry(Qgis.GeometryValidationEngine(method), flags)) File "C:\OSGeo4W/apps/qgis/./python/plugins\processing\algs\qgis\CheckValidity.py", line 124 in processAlgorithm return self.doCheck( Thread 0x00023220 (most recent call first): File "C:\OSGeo4W/apps/qgis/./python/plugins\processing\ProcessingPlugin.py", line 395 in executeAlgorithm dlg.exec_() File "C:\OSGeo4W/apps/qgis/./python/plugins\processing\gui\ProcessingToolbox.py", line 234 in executeAlgorithm self.executeWithGui.emit(alg.id(), self, self.in_place_mode, False) File "C:\OSGeo4W/apps/qgis/./python/plugins\processing\gui\AlgorithmLocatorFilter.py", line 120 in triggerResult dlg.exec_() ``` **Stack Trace** No stack trace is available. **QGIS Info** QGIS Version: 3.24.2-Tisler QGIS code revision: 13c1a028 Compiled against Qt: 5.15.2 Running against Qt: 5.15.2 Compiled against GDAL: 3.4.2 Running against GDAL: 3.4.2 **System Info** CPU Type: x86_64 Kernel Type: winnt Kernel Version: 10.0.19044 ### Steps to reproduce the issue 1. Open this geopackage [checkvalidity_crash.zip](https://github.com/qgis/QGIS/files/8663128/checkvalidity_crash.zip) in QGIS 2. Run the processing tool "Check Validity" with method = QGIS 3. Crash! ### Versions QGIS version 3.24.2-Tisler QGIS code revision 13c1a028 Qt version 5.15.2 Python version 3.9.5 GDAL/OGR version 3.4.2 PROJ version 9.0.0 EPSG Registry database version v10.054 (2022-02-13) GEOS version 3.10.2-CAPI-1.16.0 SQLite version 3.38.1 PDAL version 2.3.0 PostgreSQL client version unknown SpatiaLite version 5.0.1 QWT version 6.1.3 QScintilla2 version 2.11.5 OS version Windows 10 Version 2009 Active Python plugins db_manager 0.1.20 grassprovider 2.12.99 MetaSearch 0.3.6 processing 2.12.99 sagaprovider 2.12.99 ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [X] I tried with a new QGIS profile ### Additional context _No response_
1.0
Crash when running check validity using QGIS method - ### What is the bug or the crash? When running "Check Validity" on a particular geopackage using the QGIS method, QGIS crashes. When using the GEOS method, it reports no errors. I've also tested on a clean user profile, and the crash still occurs. I've isolated the problem to this single multi-polygon feature in a geopackage: [checkvalidity_crash.zip](https://github.com/qgis/QGIS/files/8663128/checkvalidity_crash.zip) ## Report Details **Python Stack Trace** ``` Windows fatal exception: access violation Current thread 0x00024620 (most recent call first): File "C:\OSGeo4W/apps/qgis/./python/plugins\processing\algs\qgis\CheckValidity.py", line 160 in doCheck errors = list(geom.validateGeometry(Qgis.GeometryValidationEngine(method), flags)) File "C:\OSGeo4W/apps/qgis/./python/plugins\processing\algs\qgis\CheckValidity.py", line 124 in processAlgorithm return self.doCheck( Thread 0x00023220 (most recent call first): File "C:\OSGeo4W/apps/qgis/./python/plugins\processing\ProcessingPlugin.py", line 395 in executeAlgorithm dlg.exec_() File "C:\OSGeo4W/apps/qgis/./python/plugins\processing\gui\ProcessingToolbox.py", line 234 in executeAlgorithm self.executeWithGui.emit(alg.id(), self, self.in_place_mode, False) File "C:\OSGeo4W/apps/qgis/./python/plugins\processing\gui\AlgorithmLocatorFilter.py", line 120 in triggerResult dlg.exec_() ``` **Stack Trace** No stack trace is available. **QGIS Info** QGIS Version: 3.24.2-Tisler QGIS code revision: 13c1a028 Compiled against Qt: 5.15.2 Running against Qt: 5.15.2 Compiled against GDAL: 3.4.2 Running against GDAL: 3.4.2 **System Info** CPU Type: x86_64 Kernel Type: winnt Kernel Version: 10.0.19044 ### Steps to reproduce the issue 1. Open this geopackage [checkvalidity_crash.zip](https://github.com/qgis/QGIS/files/8663128/checkvalidity_crash.zip) in QGIS 2. Run the processing tool "Check Validity" with method = QGIS 3. Crash! ### Versions QGIS version 3.24.2-Tisler QGIS code revision 13c1a028 Qt version 5.15.2 Python version 3.9.5 GDAL/OGR version 3.4.2 PROJ version 9.0.0 EPSG Registry database version v10.054 (2022-02-13) GEOS version 3.10.2-CAPI-1.16.0 SQLite version 3.38.1 PDAL version 2.3.0 PostgreSQL client version unknown SpatiaLite version 5.0.1 QWT version 6.1.3 QScintilla2 version 2.11.5 OS version Windows 10 Version 2009 Active Python plugins db_manager 0.1.20 grassprovider 2.12.99 MetaSearch 0.3.6 processing 2.12.99 sagaprovider 2.12.99 ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [X] I tried with a new QGIS profile ### Additional context _No response_
process
crash when running check validity using qgis method what is the bug or the crash when running check validity on a particular geopackage using the qgis method qgis crashes when using the geos method it reports no errors i ve also tested on a clean user profile and the crash still occurs i ve isolated the problem to this single multi polygon feature in a geopackage report details python stack trace windows fatal exception access violation current thread most recent call first file c apps qgis python plugins processing algs qgis checkvalidity py line in docheck errors list geom validategeometry qgis geometryvalidationengine method flags file c apps qgis python plugins processing algs qgis checkvalidity py line in processalgorithm return self docheck thread most recent call first file c apps qgis python plugins processing processingplugin py line in executealgorithm dlg exec file c apps qgis python plugins processing gui processingtoolbox py line in executealgorithm self executewithgui emit alg id self self in place mode false file c apps qgis python plugins processing gui algorithmlocatorfilter py line in triggerresult dlg exec stack trace no stack trace is available qgis info qgis version tisler qgis code revision compiled against qt running against qt compiled against gdal running against gdal system info cpu type kernel type winnt kernel version steps to reproduce the issue open this geopackage in qgis run the processing tool check validity with method qgis crash versions qgis version tisler qgis code revision qt version python version gdal ogr version proj version epsg registry database version geos version capi sqlite version pdal version postgresql client version unknown spatialite version qwt version version os version windows version active python plugins db manager grassprovider metasearch processing sagaprovider supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context no response
1
19,706
26,053,298,060
IssuesEvent
2022-12-22 21:16:39
MPMG-DCC-UFMG/C01
https://api.github.com/repos/MPMG-DCC-UFMG/C01
opened
Interface de passos com Vue.js - Tratamento de contextos
[1] Bug [0] Desenvolvimento [2] Média Prioridade [3] Processamento Dinâmico
## Comportamento Esperado Um passo dentro do contexto de nova aba ou iframe deve possuir uma cor diferente para indicar esse contexto, conforme era feito na versão anterior da interface de passos. Além disso, deve ser tratado o caso onde uma alteração na interface invalide o contexto de um passo, por exemplo: se há um passo de "fechar aba" após um "abrir em nova aba", e esse passo de "abrir em nova aba" é removido, algo deve ser feito para manter a consistência da configuração (remover o "fechar aba" correspondente, exibir uma mensagem para o usuário indicando da situação, impedir o salvamento do coletor até a correção do problema, etc). Também devem ser tratadas as situações onde o passo que altera o contexto é editado, movido, etc. ## Comportamento Atual Nenhum tratamento especial é dado aos passos devido ao contexto de execução. ## Sistema Branch `issue-882`.
1.0
Interface de passos com Vue.js - Tratamento de contextos - ## Comportamento Esperado Um passo dentro do contexto de nova aba ou iframe deve possuir uma cor diferente para indicar esse contexto, conforme era feito na versão anterior da interface de passos. Além disso, deve ser tratado o caso onde uma alteração na interface invalide o contexto de um passo, por exemplo: se há um passo de "fechar aba" após um "abrir em nova aba", e esse passo de "abrir em nova aba" é removido, algo deve ser feito para manter a consistência da configuração (remover o "fechar aba" correspondente, exibir uma mensagem para o usuário indicando da situação, impedir o salvamento do coletor até a correção do problema, etc). Também devem ser tratadas as situações onde o passo que altera o contexto é editado, movido, etc. ## Comportamento Atual Nenhum tratamento especial é dado aos passos devido ao contexto de execução. ## Sistema Branch `issue-882`.
process
interface de passos com vue js tratamento de contextos comportamento esperado um passo dentro do contexto de nova aba ou iframe deve possuir uma cor diferente para indicar esse contexto conforme era feito na versão anterior da interface de passos além disso deve ser tratado o caso onde uma alteração na interface invalide o contexto de um passo por exemplo se há um passo de fechar aba após um abrir em nova aba e esse passo de abrir em nova aba é removido algo deve ser feito para manter a consistência da configuração remover o fechar aba correspondente exibir uma mensagem para o usuário indicando da situação impedir o salvamento do coletor até a correção do problema etc também devem ser tratadas as situações onde o passo que altera o contexto é editado movido etc comportamento atual nenhum tratamento especial é dado aos passos devido ao contexto de execução sistema branch issue
1
49,161
6,150,141,237
IssuesEvent
2017-06-27 21:45:35
navx2810/gbs-grm
https://api.github.com/repos/navx2810/gbs-grm
opened
Create flow of application.
design
The flow is the "flow" of the application from page/component to page/component. Usually done in a diagram.
1.0
Create flow of application. - The flow is the "flow" of the application from page/component to page/component. Usually done in a diagram.
non_process
create flow of application the flow is the flow of the application from page component to page component usually done in a diagram
0
7,903
3,633,156,525
IssuesEvent
2016-02-11 13:28:01
catapult-project/catapult
https://api.github.com/repos/catapult-project/catapult
opened
Get rid of <tr-ui-u-time-stamp-span> and <tr-ui-u-time-duration-span>
Code Health Good First Bug
_[Follow-up for #1981]_ Both elements are unnecessary wrappers around <tt>\<tr-ui-u-scalar-span\></tt>. I propose doing the following: * Replace <tt>\<tr-ui-u-__time-stamp__-span\></tt> with <tt>\<tr-ui-u-__scalar__-span __unit="timeStampInMs"__\></tt> and * Replace <tt>tr.ui.units.create**TimeStamp**Span(x)</tt> with <tt>tr.ui.units.create**Scalar**Span(__new tr.b.u.Scalar__(x, __tr.b.u.Unit.byName.timeStampInMs__))</tt>. @natduca: Sounds good?
1.0
Get rid of <tr-ui-u-time-stamp-span> and <tr-ui-u-time-duration-span> - _[Follow-up for #1981]_ Both elements are unnecessary wrappers around <tt>\<tr-ui-u-scalar-span\></tt>. I propose doing the following: * Replace <tt>\<tr-ui-u-__time-stamp__-span\></tt> with <tt>\<tr-ui-u-__scalar__-span __unit="timeStampInMs"__\></tt> and * Replace <tt>tr.ui.units.create**TimeStamp**Span(x)</tt> with <tt>tr.ui.units.create**Scalar**Span(__new tr.b.u.Scalar__(x, __tr.b.u.Unit.byName.timeStampInMs__))</tt>. @natduca: Sounds good?
non_process
get rid of and both elements are unnecessary wrappers around i propose doing the following replace with and replace tr ui units create timestamp span x with tr ui units create scalar span new tr b u scalar x tr b u unit byname timestampinms natduca sounds good
0
4,040
6,972,783,515
IssuesEvent
2017-12-11 18:10:45
triplea-game/triplea
https://api.github.com/repos/triplea-game/triplea
reopened
Move install4j bundled JREs from GitHub to Linode
category: dev & admin process discussion type: process
We bundle JREs with our installer for users that do not have a Java 8 JRE installed on their machine. We currently host these bundled JREs in the triplea-game/assets repo, and the installer downloads them, if needed, directly from GitHub. There are a few problems hosting these files on GitHub: * We can't just provide the "latest" JRE because older installers are built with a hard-coded link to a specific path in the repo. As long as those older builds are considered compatible, we have to keep multiple JRE versions on `HEAD` in the repo (we currently have two versions for three platforms). Each JRE is approximately 350 MiB. * Adding additional JREs increases the time to run the Gradle build on Travis due to the additional download requirements. * Adding additional JREs increases the size of the triplea-game/assets repo significantly. Even when we can remove an older bundled JRE from `HEAD`, it stays in the repo history unless we rewrite history. A static file server is probably a better place to host these resources rather than using a Git repo. **The purpose of this issue is to discuss possibly moving the bundled JREs to one of our Linode servers and serve them from there using Nginx (or something equivalent).** Some issues with self-hosting include: 1. Increased bandwidth possibly leading to an increased Linode bill. It would be great if we could get some metrics from GitHub to see how often the bundled JREs are downloaded so we can predict how much traffic we'll see. 1. The host should be available as much as possible. It should not be taken down except for maintenance. 1. The host name should not change over time. We won't be able to go back and modify installers from older releases to point them to a new host. I'm not familiar with this aspect of our Linode setup, so if changing host names is common, we might have to consider having a reverse proxy or something to redirect from the old name to the new name. @DanVanAtta @prastle @RoiEXLab @ron-murhammer Thoughts?
2.0
Move install4j bundled JREs from GitHub to Linode - We bundle JREs with our installer for users that do not have a Java 8 JRE installed on their machine. We currently host these bundled JREs in the triplea-game/assets repo, and the installer downloads them, if needed, directly from GitHub. There are a few problems hosting these files on GitHub: * We can't just provide the "latest" JRE because older installers are built with a hard-coded link to a specific path in the repo. As long as those older builds are considered compatible, we have to keep multiple JRE versions on `HEAD` in the repo (we currently have two versions for three platforms). Each JRE is approximately 350 MiB. * Adding additional JREs increases the time to run the Gradle build on Travis due to the additional download requirements. * Adding additional JREs increases the size of the triplea-game/assets repo significantly. Even when we can remove an older bundled JRE from `HEAD`, it stays in the repo history unless we rewrite history. A static file server is probably a better place to host these resources rather than using a Git repo. **The purpose of this issue is to discuss possibly moving the bundled JREs to one of our Linode servers and serve them from there using Nginx (or something equivalent).** Some issues with self-hosting include: 1. Increased bandwidth possibly leading to an increased Linode bill. It would be great if we could get some metrics from GitHub to see how often the bundled JREs are downloaded so we can predict how much traffic we'll see. 1. The host should be available as much as possible. It should not be taken down except for maintenance. 1. The host name should not change over time. We won't be able to go back and modify installers from older releases to point them to a new host. I'm not familiar with this aspect of our Linode setup, so if changing host names is common, we might have to consider having a reverse proxy or something to redirect from the old name to the new name. @DanVanAtta @prastle @RoiEXLab @ron-murhammer Thoughts?
process
move bundled jres from github to linode we bundle jres with our installer for users that do not have a java jre installed on their machine we currently host these bundled jres in the triplea game assets repo and the installer downloads them if needed directly from github there are a few problems hosting these files on github we can t just provide the latest jre because older installers are built with a hard coded link to a specific path in the repo as long as those older builds are considered compatible we have to keep multiple jre versions on head in the repo we currently have two versions for three platforms each jre is approximately mib adding additional jres increases the time to run the gradle build on travis due to the additional download requirements adding additional jres increases the size of the triplea game assets repo significantly even when we can remove an older bundled jre from head it stays in the repo history unless we rewrite history a static file server is probably a better place to host these resources rather than using a git repo the purpose of this issue is to discuss possibly moving the bundled jres to one of our linode servers and serve them from there using nginx or something equivalent some issues with self hosting include increased bandwidth possibly leading to an increased linode bill it would be great if we could get some metrics from github to see how often the bundled jres are downloaded so we can predict how much traffic we ll see the host should be available as much as possible it should not be taken down except for maintenance the host name should not change over time we won t be able to go back and modify installers from older releases to point them to a new host i m not familiar with this aspect of our linode setup so if changing host names is common we might have to consider having a reverse proxy or something to redirect from the old name to the new name danvanatta prastle roiexlab ron murhammer thoughts
1
2,775
5,712,694,876
IssuesEvent
2017-04-19 04:46:45
kerubistan/kerub
https://api.github.com/repos/kerubistan/kerub
opened
nulls injected through json
bug component: security component:data processing priority: high
while nulls are not expected on the server, they can be injected through json https://twitter.com/kozka/status/854439819216396288 this would be great if jackson could handle this, but it is "wontfix" https://github.com/FasterXML/jackson-module-kotlin/issues/27
1.0
nulls injected through json - while nulls are not expected on the server, they can be injected through json https://twitter.com/kozka/status/854439819216396288 this would be great if jackson could handle this, but it is "wontfix" https://github.com/FasterXML/jackson-module-kotlin/issues/27
process
nulls injected through json while nulls are not expected on the server they can be injected through json this would be great if jackson could handle this but it is wontfix
1
22,218
30,768,981,159
IssuesEvent
2023-07-30 17:06:41
km4ack/73Linux
https://api.github.com/repos/km4ack/73Linux
closed
73Linux/x86LMint Update Issues
in process
So, earlier I messaged about problems with getting the IC-705 to work with WSJT-X. The fix appeared to be to update WSJT-X to the latest revision, which uses an updated HAMLIB that does appear to support the IC-705. Unfortunately, noticing that I was unable to update WSJT-X (using the 73Linux update tool), I tried to update several other installed applications one at a time, including CHIRP, HAMLIB and FLRIG, without success. The update operation completed to the "REBOOT" dialog and upon reboot the requested application was not updated. There was no notice-able error presented in either the on-screen log or the stored on in ~/73linux/cache/logs, but no update. Gateway x86-64 laptop with LMint installed under 73Linux; plenty of storage and memory. Any thoughts? Thanks, -- Jeff Marden N1JCM
1.0
73Linux/x86LMint Update Issues - So, earlier I messaged about problems with getting the IC-705 to work with WSJT-X. The fix appeared to be to update WSJT-X to the latest revision, which uses an updated HAMLIB that does appear to support the IC-705. Unfortunately, noticing that I was unable to update WSJT-X (using the 73Linux update tool), I tried to update several other installed applications one at a time, including CHIRP, HAMLIB and FLRIG, without success. The update operation completed to the "REBOOT" dialog and upon reboot the requested application was not updated. There was no notice-able error presented in either the on-screen log or the stored on in ~/73linux/cache/logs, but no update. Gateway x86-64 laptop with LMint installed under 73Linux; plenty of storage and memory. Any thoughts? Thanks, -- Jeff Marden N1JCM
process
update issues so earlier i messaged about problems with getting the ic to work with wsjt x the fix appeared to be to update wsjt x to the latest revision which uses an updated hamlib that does appear to support the ic unfortunately noticing that i was unable to update wsjt x using the update tool i tried to update several other installed applications one at a time including chirp hamlib and flrig without success the update operation completed to the reboot dialog and upon reboot the requested application was not updated there was no notice able error presented in either the on screen log or the stored on in cache logs but no update gateway laptop with lmint installed under plenty of storage and memory any thoughts thanks jeff marden
1
220,229
16,902,865,065
IssuesEvent
2021-06-24 00:58:54
bethlakshmi/gbe-divio-djangocms-python2.7
https://api.github.com/repos/bethlakshmi/gbe-divio-djangocms-python2.7
opened
Wrong URL in ticketing page doc
bug documentation
If you go to edit a ticket item there's some help test next to the field where you can indicate which "simple Icon" is displayed. The URL given is wrong and should be https://simplelineicons.github.io/
1.0
Wrong URL in ticketing page doc - If you go to edit a ticket item there's some help test next to the field where you can indicate which "simple Icon" is displayed. The URL given is wrong and should be https://simplelineicons.github.io/
non_process
wrong url in ticketing page doc if you go to edit a ticket item there s some help test next to the field where you can indicate which simple icon is displayed the url given is wrong and should be
0
251,129
8,000,370,873
IssuesEvent
2018-07-22 15:11:18
krshubham/interview-prep
https://api.github.com/repos/krshubham/interview-prep
closed
App Tooling Upgrade
Priority:low
## DB Move to RethinkDB over MongoDB. It's perfect for our usage. https://rethinkdb.com/faq/ ## Auth Use Passport.js (when we move to the model where users can sign in and reply) ## API Move to GraphQl (Apollo toolchain) ##Express Setup it like it's here: https://github.com/withspectrum/spectrum/blob/alpha/api/index.js ## Javascript Tooling There is a lot of work that's been done in the JS ecosystem to finally put the fires down between developers. Why not use them? Our team is primarily using VSCode as the editor, but feel free to use any which supports these tools (most popular editor have extensions for them): * Prettier. * Eslint * Flow (will introduce when code has to scale) Eslint * Too many console.logs spoil the broth. Warning for console.logs * Introduce Airbnb's eslint config. Good extensions to have: https://marketplace.visualstudio.com/items?itemName=burkeholland.simple-react-snippets
1.0
App Tooling Upgrade - ## DB Move to RethinkDB over MongoDB. It's perfect for our usage. https://rethinkdb.com/faq/ ## Auth Use Passport.js (when we move to the model where users can sign in and reply) ## API Move to GraphQl (Apollo toolchain) ##Express Setup it like it's here: https://github.com/withspectrum/spectrum/blob/alpha/api/index.js ## Javascript Tooling There is a lot of work that's been done in the JS ecosystem to finally put the fires down between developers. Why not use them? Our team is primarily using VSCode as the editor, but feel free to use any which supports these tools (most popular editor have extensions for them): * Prettier. * Eslint * Flow (will introduce when code has to scale) Eslint * Too many console.logs spoil the broth. Warning for console.logs * Introduce Airbnb's eslint config. Good extensions to have: https://marketplace.visualstudio.com/items?itemName=burkeholland.simple-react-snippets
non_process
app tooling upgrade db move to rethinkdb over mongodb it s perfect for our usage auth use passport js when we move to the model where users can sign in and reply api move to graphql apollo toolchain express setup it like it s here javascript tooling there is a lot of work that s been done in the js ecosystem to finally put the fires down between developers why not use them our team is primarily using vscode as the editor but feel free to use any which supports these tools most popular editor have extensions for them prettier eslint flow will introduce when code has to scale eslint too many console logs spoil the broth warning for console logs introduce airbnb s eslint config good extensions to have
0
22,880
20,424,097,350
IssuesEvent
2022-02-24 00:40:19
bevyengine/bevy
https://api.github.com/repos/bevyengine/bevy
closed
#[derive(Query)]
C-Enhancement A-ECS C-Usability
**What problem does this solve or what need does it fill?** Giant tuples within query (`Query<(&Foo, &Bar, &mut Baz)>`) is really annoying as then everything has to be referenced by `.0`, etc, and if I decide to add `Entity` to the list I have to shift everything. **Describe the solution would you like?** Make a derive macro `#[derive(Query)]` which implements `HecsQuery` for the structure in question; as such the structure can be used as a query. **Describe the alternative(s) you've considered?** None. **Additional context** This could also support `Or` by allowing deriving an enum.
True
#[derive(Query)] - **What problem does this solve or what need does it fill?** Giant tuples within query (`Query<(&Foo, &Bar, &mut Baz)>`) is really annoying as then everything has to be referenced by `.0`, etc, and if I decide to add `Entity` to the list I have to shift everything. **Describe the solution would you like?** Make a derive macro `#[derive(Query)]` which implements `HecsQuery` for the structure in question; as such the structure can be used as a query. **Describe the alternative(s) you've considered?** None. **Additional context** This could also support `Or` by allowing deriving an enum.
non_process
what problem does this solve or what need does it fill giant tuples within query query is really annoying as then everything has to be referenced by etc and if i decide to add entity to the list i have to shift everything describe the solution would you like make a derive macro which implements hecsquery for the structure in question as such the structure can be used as a query describe the alternative s you ve considered none additional context this could also support or by allowing deriving an enum
0
20,687
27,358,213,024
IssuesEvent
2023-02-27 14:16:31
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Is the listed Throw syntax correct?
automation/svc triaged cxp doc-enhancement process-automation/subsvc Pri2
[Enter feedback here] I think the Throw syntax in the screenshot in learn is incorrect. Is this correct? ![image](https://user-images.githubusercontent.com/124872654/220248692-91b65a10-ae1b-4229-b001-5947215c6862.png) --- #### Document Details ⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.* * ID: 8ab9bff0-9179-827d-1a8d-c58a98625070 * Version Independent ID: bf46f240-6c29-f040-6b0f-b454606b04a1 * Content: [Handle errors in Azure Automation graphical runbooks](https://learn.microsoft.com/en-us/azure/automation/automation-runbook-graphical-error-handling#turn-exceptions-into-non-terminating-errors) * Content Source: [articles/automation/automation-runbook-graphical-error-handling.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/automation/automation-runbook-graphical-error-handling.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @SnehaSudhirG * Microsoft Alias: **sudhirsneha**
1.0
Is the listed Throw syntax correct? - [Enter feedback here] I think the Throw syntax in the screenshot in learn is incorrect. Is this correct? ![image](https://user-images.githubusercontent.com/124872654/220248692-91b65a10-ae1b-4229-b001-5947215c6862.png) --- #### Document Details ⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.* * ID: 8ab9bff0-9179-827d-1a8d-c58a98625070 * Version Independent ID: bf46f240-6c29-f040-6b0f-b454606b04a1 * Content: [Handle errors in Azure Automation graphical runbooks](https://learn.microsoft.com/en-us/azure/automation/automation-runbook-graphical-error-handling#turn-exceptions-into-non-terminating-errors) * Content Source: [articles/automation/automation-runbook-graphical-error-handling.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/automation/automation-runbook-graphical-error-handling.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @SnehaSudhirG * Microsoft Alias: **sudhirsneha**
process
is the listed throw syntax correct i think the throw syntax in the screenshot in learn is incorrect is this correct document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login snehasudhirg microsoft alias sudhirsneha
1
12,054
14,739,179,187
IssuesEvent
2021-01-07 06:39:48
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
'No records found' should be displayed when admin doesn't have any site level, study level, app level permission
Bug P1 Participant manager Process: Fixed Process: Tested QA Process: Tested dev
1. 'No records found' text should be displayed when there are no data in the database 2. 'No records found' text should be displayed when there are some data in the database and the user does not have any permission for sites 3. Proper custom messages should be displayed when a normal admin user doesn't meet conditions related to permission for the following i. Study level permission AR : 'This view displays study-wise enrollment if you manage multiple sites' error message is displayed ER : 'This view displays study-wise enrollment if you manage multiple sites' text should be displayed in the attached screen instead of an error message ii. App-level permission AR : 'This view displays app-wise enrollment if you manage multiple studies.' error message is displayed ER : 'This view displays app-wise enrollment if you manage multiple studies' text should be displayed in the attached screen instead of an error message ER: ![permission ui](https://user-images.githubusercontent.com/71445210/103070664-fe3b6780-45e7-11eb-98d1-6e81689fdce8.png)
3.0
'No records found' should be displayed when admin doesn't have any site level, study level, app level permission - 1. 'No records found' text should be displayed when there are no data in the database 2. 'No records found' text should be displayed when there are some data in the database and the user does not have any permission for sites 3. Proper custom messages should be displayed when a normal admin user doesn't meet conditions related to permission for the following i. Study level permission AR : 'This view displays study-wise enrollment if you manage multiple sites' error message is displayed ER : 'This view displays study-wise enrollment if you manage multiple sites' text should be displayed in the attached screen instead of an error message ii. App-level permission AR : 'This view displays app-wise enrollment if you manage multiple studies.' error message is displayed ER : 'This view displays app-wise enrollment if you manage multiple studies' text should be displayed in the attached screen instead of an error message ER: ![permission ui](https://user-images.githubusercontent.com/71445210/103070664-fe3b6780-45e7-11eb-98d1-6e81689fdce8.png)
process
no records found should be displayed when admin doesn t have any site level study level app level permission no records found text should be displayed when there are no data in the database no records found text should be displayed when there are some data in the database and the user does not have any permission for sites proper custom messages should be displayed when a normal admin user doesn t meet conditions related to permission for the following i study level permission ar this view displays study wise enrollment if you manage multiple sites error message is displayed er this view displays study wise enrollment if you manage multiple sites text should be displayed in the attached screen instead of an error message ii app level permission ar this view displays app wise enrollment if you manage multiple studies error message is displayed er this view displays app wise enrollment if you manage multiple studies text should be displayed in the attached screen instead of an error message er
1
9,162
3,258,419,954
IssuesEvent
2015-10-20 22:15:35
radical-cybertools/radical.pilot
https://api.github.com/repos/radical-cybertools/radical.pilot
closed
a better title?
documentation
4.2 "Obtaining Unit Details" --> 4.2 "Inspecting Execution" or something to that effect?
1.0
a better title? - 4.2 "Obtaining Unit Details" --> 4.2 "Inspecting Execution" or something to that effect?
non_process
a better title obtaining unit details inspecting execution or something to that effect
0
19,470
25,767,606,786
IssuesEvent
2022-12-09 04:07:00
dtcenter/MET
https://api.github.com/repos/dtcenter/MET
closed
Fix logic in reading AERONET v3 data
type: bug component: user support requestor: NOAA/EMC reporting: DTC NOAA R2O required: FOR OFFICIAL RELEASE MET: PreProcessing Tools (Point) priority: high
## Describe the Problem ## This issue arose with the METplus Discussion [dtcenter/METplus#1888](https://github.com/dtcenter/METplus/discussions/1888). When the user obtained AERONET version 3 data from the [official site](https://aeronet.gsfc.nasa.gov) it was determined that the format was not consistent with the format of the data we currently use for our [unit testing](https://dtcenter.ucar.edu/dfiles/code/METplus/MET/MET_unit_test/unit_test/obs_data/aeronet/20150917_20150926_Table_Mountain.lev20). Sample files from the new format and @hsoh-u's analysis of what would need to be done for this issue can be viewed within the Discussion at this [post](https://github.com/dtcenter/METplus/discussions/1888#discussioncomment-4319771). ### Expected Behavior ### The MET code should be able to read AERONET v3 data obtained from the official site. ### Environment ### Describe your runtime environment: *1. Machine: not specific* *2. OS: not specific* *3. Software version number: all current releases* ### To Reproduce ### Describe the steps to reproduce the behavior: *1. Go to '...'* *2. Click on '....'* *3. Scroll down to '....'* *4. See error* *Post relevant sample data following these instructions:* *https://dtcenter.org/community-code/model-evaluation-tools-met/met-help-desk#ftp* Review the log file from Partha and read the data files he provided. ### Relevant Deadlines ### MET-11.0.0 Official Release (12/7 - 12/9) ### Funding Source ### 2773542 ## Define the Metadata ## ### Assignee ### - [x] Select **engineer(s)** or **no engineer** required - [ ] Select **scientist(s)** or **no scientist** required ### Labels ### - [x] Select **component(s)** - [x] Select **priority** - [x] Select **requestor(s)** ### Projects and Milestone ### - [x] Select **Organization** level **Project** for support of the current coordinated release - [x] Select **Repository** level **Project** for development toward the next official release or add **alert: NEED PROJECT ASSIGNMENT** label - [x] Select **Milestone** as the next bugfix version ## Define Related Issue(s) ## Consider the impact to the other METplus components. - [x] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdataio](https://github.com/dtcenter/METdataio/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose) ## Bugfix Checklist ## See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details. - [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**. - [ ] Fork this repository or create a branch of **main_\<Version>**. Branch name: `bugfix_<Issue Number>_main_<Version>_<Description>` - [ ] Fix the bug and test your changes. - [ ] Add/update log messages for easier debugging. - [ ] Add/update unit tests. - [ ] Add/update documentation. - [ ] Push local changes to GitHub. - [ ] Submit a pull request to merge into **main_\<Version>**. Pull request: `bugfix <Issue Number> main_<Version> <Description>` - [ ] Define the pull request metadata, as permissions allow. Select: **Reviewer(s)** and **Linked issues** Select: **Organization** level software support **Project** for the current coordinated release Select: **Milestone** as the next bugfix version - [ ] Iterate until the reviewer(s) accept and merge your changes. - [ ] Delete your fork or branch. - [ ] Complete the steps above to fix the bug on the **develop** branch. Branch name: `bugfix_<Issue Number>_develop_<Description>` Pull request: `bugfix <Issue Number> develop <Description>` Select: **Reviewer(s)** and **Linked issues** Select: **Repository** level development cycle **Project** for the next official release Select: **Milestone** as the next official version - [ ] Close this issue.
1.0
Fix logic in reading AERONET v3 data - ## Describe the Problem ## This issue arose with the METplus Discussion [dtcenter/METplus#1888](https://github.com/dtcenter/METplus/discussions/1888). When the user obtained AERONET version 3 data from the [official site](https://aeronet.gsfc.nasa.gov) it was determined that the format was not consistent with the format of the data we currently use for our [unit testing](https://dtcenter.ucar.edu/dfiles/code/METplus/MET/MET_unit_test/unit_test/obs_data/aeronet/20150917_20150926_Table_Mountain.lev20). Sample files from the new format and @hsoh-u's analysis of what would need to be done for this issue can be viewed within the Discussion at this [post](https://github.com/dtcenter/METplus/discussions/1888#discussioncomment-4319771). ### Expected Behavior ### The MET code should be able to read AERONET v3 data obtained from the official site. ### Environment ### Describe your runtime environment: *1. Machine: not specific* *2. OS: not specific* *3. Software version number: all current releases* ### To Reproduce ### Describe the steps to reproduce the behavior: *1. Go to '...'* *2. Click on '....'* *3. Scroll down to '....'* *4. See error* *Post relevant sample data following these instructions:* *https://dtcenter.org/community-code/model-evaluation-tools-met/met-help-desk#ftp* Review the log file from Partha and read the data files he provided. ### Relevant Deadlines ### MET-11.0.0 Official Release (12/7 - 12/9) ### Funding Source ### 2773542 ## Define the Metadata ## ### Assignee ### - [x] Select **engineer(s)** or **no engineer** required - [ ] Select **scientist(s)** or **no scientist** required ### Labels ### - [x] Select **component(s)** - [x] Select **priority** - [x] Select **requestor(s)** ### Projects and Milestone ### - [x] Select **Organization** level **Project** for support of the current coordinated release - [x] Select **Repository** level **Project** for development toward the next official release or add **alert: NEED PROJECT ASSIGNMENT** label - [x] Select **Milestone** as the next bugfix version ## Define Related Issue(s) ## Consider the impact to the other METplus components. - [x] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdataio](https://github.com/dtcenter/METdataio/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose) ## Bugfix Checklist ## See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details. - [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**. - [ ] Fork this repository or create a branch of **main_\<Version>**. Branch name: `bugfix_<Issue Number>_main_<Version>_<Description>` - [ ] Fix the bug and test your changes. - [ ] Add/update log messages for easier debugging. - [ ] Add/update unit tests. - [ ] Add/update documentation. - [ ] Push local changes to GitHub. - [ ] Submit a pull request to merge into **main_\<Version>**. Pull request: `bugfix <Issue Number> main_<Version> <Description>` - [ ] Define the pull request metadata, as permissions allow. Select: **Reviewer(s)** and **Linked issues** Select: **Organization** level software support **Project** for the current coordinated release Select: **Milestone** as the next bugfix version - [ ] Iterate until the reviewer(s) accept and merge your changes. - [ ] Delete your fork or branch. - [ ] Complete the steps above to fix the bug on the **develop** branch. Branch name: `bugfix_<Issue Number>_develop_<Description>` Pull request: `bugfix <Issue Number> develop <Description>` Select: **Reviewer(s)** and **Linked issues** Select: **Repository** level development cycle **Project** for the next official release Select: **Milestone** as the next official version - [ ] Close this issue.
process
fix logic in reading aeronet data describe the problem this issue arose with the metplus discussion when the user obtained aeronet version data from the it was determined that the format was not consistent with the format of the data we currently use for our sample files from the new format and hsoh u s analysis of what would need to be done for this issue can be viewed within the discussion at this expected behavior the met code should be able to read aeronet data obtained from the official site environment describe your runtime environment machine not specific os not specific software version number all current releases to reproduce describe the steps to reproduce the behavior go to click on scroll down to see error post relevant sample data following these instructions review the log file from partha and read the data files he provided relevant deadlines met official release funding source define the metadata assignee select engineer s or no engineer required select scientist s or no scientist required labels select component s select priority select requestor s projects and milestone select organization level project for support of the current coordinated release select repository level project for development toward the next official release or add alert need project assignment label select milestone as the next bugfix version define related issue s consider the impact to the other metplus components bugfix checklist see the for details complete the issue definition above including the time estimate and funding source fork this repository or create a branch of main branch name bugfix main fix the bug and test your changes add update log messages for easier debugging add update unit tests add update documentation push local changes to github submit a pull request to merge into main pull request bugfix main define the pull request metadata as permissions allow select reviewer s and linked issues select organization level software support project for the current coordinated release select milestone as the next bugfix version iterate until the reviewer s accept and merge your changes delete your fork or branch complete the steps above to fix the bug on the develop branch branch name bugfix develop pull request bugfix develop select reviewer s and linked issues select repository level development cycle project for the next official release select milestone as the next official version close this issue
1
13,135
15,555,305,893
IssuesEvent
2021-03-16 05:51:59
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
gdal_rasterize update
Feature Request Processing
Hello everyone, according to gdal_rasterize documentation it is possible to use -3d option in order to use the Z attribute to rasterize. Is it possible to include that feature in QGIS processing? I tried, as a test, to include the "-3d" in advanced parameters, but an error is thrown because the parameter "burn value" (claimed as optional) is anyway called in the GDAL string (while the filed attribute can be skipped): ``` gdal_rasterize -l INPUT -burn 0.0 -tr 5.0 5.0 -a_nodata 0.0 -te 2309012.5 4638817.5 2312552.5 4641697.5 -ot Float32 -of GTiff -co COMPRESS=DEFLATE -co PREDICTOR=2 -co ZLEVEL=9 -3d C:/Users/dtalledo/AppData/Local/Temp/processing_jPHwjv/856fc12d75c345d780105fcae4220de7/INPUT.gpkg C:/Users/dtalledo/AppData/Local/Temp/processing_jPHwjv/90b067e3959643b9b1093a3eb5c8666f/OUTPUT.tif Risultato comando GDAL: ERROR 6: One and only one of -3d, -burn or -a is required. Usage: gdal_rasterize [-b band]* [-i] [-at] {[-burn value]* | [-a attribute_name] | [-3d]} [-add] [-l layername]* [-where expression] [-sql select_statement] [-dialect dialect] [-of format] [-a_srs srs_def] [-to "NAME=VALUE"]* [-co "NAME=VALUE"]* [-a_nodata value] [-init value]* [-te xmin ymin xmax ymax] [-tr xres yres] [-tap] [-ts width height] [-ot {Byte/Int16/UInt16/UInt32/Int32/Float32/Float64/ CInt16/CInt32/CFloat32/CFloat64}] [-optim {[AUTO]/VECTOR/RASTER}] [-q] <src_datasource> <dst_filename> Il processo ha restituito un codice di errore 1 ```
1.0
gdal_rasterize update - Hello everyone, according to gdal_rasterize documentation it is possible to use -3d option in order to use the Z attribute to rasterize. Is it possible to include that feature in QGIS processing? I tried, as a test, to include the "-3d" in advanced parameters, but an error is thrown because the parameter "burn value" (claimed as optional) is anyway called in the GDAL string (while the filed attribute can be skipped): ``` gdal_rasterize -l INPUT -burn 0.0 -tr 5.0 5.0 -a_nodata 0.0 -te 2309012.5 4638817.5 2312552.5 4641697.5 -ot Float32 -of GTiff -co COMPRESS=DEFLATE -co PREDICTOR=2 -co ZLEVEL=9 -3d C:/Users/dtalledo/AppData/Local/Temp/processing_jPHwjv/856fc12d75c345d780105fcae4220de7/INPUT.gpkg C:/Users/dtalledo/AppData/Local/Temp/processing_jPHwjv/90b067e3959643b9b1093a3eb5c8666f/OUTPUT.tif Risultato comando GDAL: ERROR 6: One and only one of -3d, -burn or -a is required. Usage: gdal_rasterize [-b band]* [-i] [-at] {[-burn value]* | [-a attribute_name] | [-3d]} [-add] [-l layername]* [-where expression] [-sql select_statement] [-dialect dialect] [-of format] [-a_srs srs_def] [-to "NAME=VALUE"]* [-co "NAME=VALUE"]* [-a_nodata value] [-init value]* [-te xmin ymin xmax ymax] [-tr xres yres] [-tap] [-ts width height] [-ot {Byte/Int16/UInt16/UInt32/Int32/Float32/Float64/ CInt16/CInt32/CFloat32/CFloat64}] [-optim {[AUTO]/VECTOR/RASTER}] [-q] <src_datasource> <dst_filename> Il processo ha restituito un codice di errore 1 ```
process
gdal rasterize update hello everyone according to gdal rasterize documentation it is possible to use option in order to use the z attribute to rasterize is it possible to include that feature in qgis processing i tried as a test to include the in advanced parameters but an error is thrown because the parameter burn value claimed as optional is anyway called in the gdal string while the filed attribute can be skipped gdal rasterize l input burn tr a nodata te ot of gtiff co compress deflate co predictor co zlevel c users dtalledo appdata local temp processing jphwjv input gpkg c users dtalledo appdata local temp processing jphwjv output tif risultato comando gdal error one and only one of burn or a is required usage gdal rasterize ot byte vector raster il processo ha restituito un codice di errore
1
4,502
7,348,891,265
IssuesEvent
2018-03-08 08:42:30
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
opened
Add more relation between algorithms help
Easy Processing
It could be nice to have a more intensive use of the "See Also" section of alg description and link them. For example, a minima, when alg A is listed in alg B see also, the inverse should be true. Which is not the case currently. But there might also be missing links.
1.0
Add more relation between algorithms help - It could be nice to have a more intensive use of the "See Also" section of alg description and link them. For example, a minima, when alg A is listed in alg B see also, the inverse should be true. Which is not the case currently. But there might also be missing links.
process
add more relation between algorithms help it could be nice to have a more intensive use of the see also section of alg description and link them for example a minima when alg a is listed in alg b see also the inverse should be true which is not the case currently but there might also be missing links
1
287,160
8,805,120,492
IssuesEvent
2018-12-26 17:36:01
strapi/strapi
https://api.github.com/repos/strapi/strapi
closed
Add support for GraphQL Apollo server tracing
pr: 🚀 New feature priority: low
- [ ] **I have created my request on the Product Board before I submitted this issue** - [x] **I have looked at all the other requests on the Product Board before I submitted this issue** (Feature Request not submitted to Product board as I plan to start playing with adding this via a PR myself.) **Please describe your feature request:** Per the following: https://github.com/apollographql/apollo-server/tree/master/packages/apollo-tracing https://github.com/apollographql/apollo-tracing I'm requesting that the Apollo server tracing be enabled so that we can see live performance metrics of a GraphQL based query to determine load and plan out/negate potential damaging query and mutation requests.
1.0
Add support for GraphQL Apollo server tracing - - [ ] **I have created my request on the Product Board before I submitted this issue** - [x] **I have looked at all the other requests on the Product Board before I submitted this issue** (Feature Request not submitted to Product board as I plan to start playing with adding this via a PR myself.) **Please describe your feature request:** Per the following: https://github.com/apollographql/apollo-server/tree/master/packages/apollo-tracing https://github.com/apollographql/apollo-tracing I'm requesting that the Apollo server tracing be enabled so that we can see live performance metrics of a GraphQL based query to determine load and plan out/negate potential damaging query and mutation requests.
non_process
add support for graphql apollo server tracing i have created my request on the product board before i submitted this issue i have looked at all the other requests on the product board before i submitted this issue feature request not submitted to product board as i plan to start playing with adding this via a pr myself please describe your feature request per the following i m requesting that the apollo server tracing be enabled so that we can see live performance metrics of a graphql based query to determine load and plan out negate potential damaging query and mutation requests
0
41
2,507,672,152
IssuesEvent
2015-01-12 19:57:50
tinkerpop/tinkerpop3
https://api.github.com/repos/tinkerpop/tinkerpop3
closed
[Proposal] has() as a step modulator.
enhancement process
Here is a big internal change that may be useful, may not. A `Step` implements `HasContainerHolder` if it stores `HasContainers`. E.g. `HasStep`, `TinkerGraphStep`, `Neo4jGraphStep`. Why not just make it such that: ```java public GraphTraversal<S,E> has(...) { if(previousStep instanceof HasContainer) previousStep.addHasContainer(new HasContainer(...)); else this.addStep(new HasStep(new HasContainer(...))); return this; } ``` Why is this cool? . GraphStep can implement `HasContainerHolder`. . There is no need for a `TinkerGraphStepStrategy`, `Neo4jGraphStepStrategy`, they simply implement `HasContainerHolder`. . `RepeatStep` can implement `HasContainerHolder`. ... see below why this is cool. ```java g.V().repeat(__.out()).until().has(label,'person') ``` In other words, repeat `out()` until you reach a person. @mbroecheler @dkuppitz @pietermartin @spmallette
1.0
[Proposal] has() as a step modulator. - Here is a big internal change that may be useful, may not. A `Step` implements `HasContainerHolder` if it stores `HasContainers`. E.g. `HasStep`, `TinkerGraphStep`, `Neo4jGraphStep`. Why not just make it such that: ```java public GraphTraversal<S,E> has(...) { if(previousStep instanceof HasContainer) previousStep.addHasContainer(new HasContainer(...)); else this.addStep(new HasStep(new HasContainer(...))); return this; } ``` Why is this cool? . GraphStep can implement `HasContainerHolder`. . There is no need for a `TinkerGraphStepStrategy`, `Neo4jGraphStepStrategy`, they simply implement `HasContainerHolder`. . `RepeatStep` can implement `HasContainerHolder`. ... see below why this is cool. ```java g.V().repeat(__.out()).until().has(label,'person') ``` In other words, repeat `out()` until you reach a person. @mbroecheler @dkuppitz @pietermartin @spmallette
process
has as a step modulator here is a big internal change that may be useful may not a step implements hascontainerholder if it stores hascontainers e g hasstep tinkergraphstep why not just make it such that java public graphtraversal has if previousstep instanceof hascontainer previousstep addhascontainer new hascontainer else this addstep new hasstep new hascontainer return this why is this cool graphstep can implement hascontainerholder there is no need for a tinkergraphstepstrategy they simply implement hascontainerholder repeatstep can implement hascontainerholder see below why this is cool java g v repeat out until has label person in other words repeat out until you reach a person mbroecheler dkuppitz pietermartin spmallette
1
11,291
14,100,076,964
IssuesEvent
2020-11-06 03:08:34
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
Retain conref and keyref information after preprocessing
feature good first issue preprocess preprocess/conref preprocess/keyref priority/medium stale
Add namespaced attributes to retain conref and keyref source information. This will allow advanced users to create e.g. review PDFs which show the source of each block. The feature should be controlled by dynamic configuration; by default the feature should be disabled.
3.0
Retain conref and keyref information after preprocessing - Add namespaced attributes to retain conref and keyref source information. This will allow advanced users to create e.g. review PDFs which show the source of each block. The feature should be controlled by dynamic configuration; by default the feature should be disabled.
process
retain conref and keyref information after preprocessing add namespaced attributes to retain conref and keyref source information this will allow advanced users to create e g review pdfs which show the source of each block the feature should be controlled by dynamic configuration by default the feature should be disabled
1
64,149
14,657,456,207
IssuesEvent
2020-12-28 15:38:13
fu1771695yongxie/yarn
https://api.github.com/repos/fu1771695yongxie/yarn
opened
CVE-2018-3750 (High) detected in io.js6ed791c665de2c1838f6080a1b377b0008cf535b
security vulnerability
## CVE-2018-3750 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>io.js6ed791c665de2c1838f6080a1b377b0008cf535b</b></p></summary> <p> <p>Node.js JavaScript runtime :sparkles::turtle::rocket::sparkles:</p> <p>Library home page: <a href=https://github.com/iojs/io.js.git>https://github.com/iojs/io.js.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/yarn/commit/b0308a6bc0041ba9a7c0fefc30be7721760dfd37">b0308a6bc0041ba9a7c0fefc30be7721760dfd37</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (0)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The utilities function in all versions <= 0.5.0 of the deep-extend node module can be tricked into modifying the prototype of Object when the attacker can control part of the structure passed to this function. This can let an attacker add or modify existing properties that will exist on all objects. <p>Publish Date: 2018-07-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3750>CVE-2018-3750</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-3750">http://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-3750</a></p> <p>Release Date: 2019-01-24</p> <p>Fix Resolution: 0.5.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-3750 (High) detected in io.js6ed791c665de2c1838f6080a1b377b0008cf535b - ## CVE-2018-3750 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>io.js6ed791c665de2c1838f6080a1b377b0008cf535b</b></p></summary> <p> <p>Node.js JavaScript runtime :sparkles::turtle::rocket::sparkles:</p> <p>Library home page: <a href=https://github.com/iojs/io.js.git>https://github.com/iojs/io.js.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/yarn/commit/b0308a6bc0041ba9a7c0fefc30be7721760dfd37">b0308a6bc0041ba9a7c0fefc30be7721760dfd37</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (0)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The utilities function in all versions <= 0.5.0 of the deep-extend node module can be tricked into modifying the prototype of Object when the attacker can control part of the structure passed to this function. This can let an attacker add or modify existing properties that will exist on all objects. <p>Publish Date: 2018-07-03 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3750>CVE-2018-3750</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-3750">http://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-3750</a></p> <p>Release Date: 2019-01-24</p> <p>Fix Resolution: 0.5.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in io cve high severity vulnerability vulnerable library io node js javascript runtime sparkles turtle rocket sparkles library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details the utilities function in all versions of the deep extend node module can be tricked into modifying the prototype of object when the attacker can control part of the structure passed to this function this can let an attacker add or modify existing properties that will exist on all objects publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
12,282
14,791,666,224
IssuesEvent
2021-01-12 13:48:24
prisma/prisma
https://api.github.com/repos/prisma/prisma
closed
A simple many to many relationship causes PANIC: 1
bug/2-confirmed kind/bug process/candidate team/client topic: broken query
## Bug description Trying to implement a simple many to many relationship results in findMany causing panic ## How to reproduce I used the following schema to create a database: ```prisma datasource db { provider = "postgresql" url = "postgresql://[snipped]" } generator client { provider = "prisma-client-js" } model Image { hash String @id tags Tag[] @relation(name: "tags") } model Tag { name String @id images Image[] @relation(name: "tags") } ``` Then lanched prisma studio and created two Image instances then saved them and created a single tag, then I associated both images with that tag. After that any prisma.image.findMany query that includes or selects images fails with a panic message (including those of prisma studio). ## Expected behavior The query should not fail since it is almost a copy of a documentation example. ```prisma //Copied from https://www.prisma.io/docs/support/help-articles/working-with-many-to-many-relations model Post { id Int @id @default(autoincrement()) title String tags Tag[] } model Tag { id Int @id @default(autoincrement()) name String @unique posts Post[] } ``` ## Prisma information Included above ## Environment & setup <!-- In which environment does the problem occur --> - OS: Linux (Manjaro) - Database: PostgreSQL - Node.js version: 15.5.0 - Prisma version: <!--[Run `prisma -v` to see your Prisma version and paste it between the ´´´]--> ``` @prisma/cli : 2.13.1 @prisma/client : 2.13.1 Current platform : debian-openssl-1.1.x Query Engine : query-engine fcbc4bb2d306c86c28014f596b1e8c7980af8bd4 (at ../../../../../home/kranga/.nvm/versions/node/v15.5.0/lib/node_modules/@prisma/cli/node_modules/@prisma/engines/query-engine-debian-openssl-1.1.x) Migration Engine : migration-engine-cli fcbc4bb2d306c86c28014f596b1e8c7980af8bd4 (at ../../../../../home/kranga/.nvm/versions/node/v15.5.0/lib/node_modules/@prisma/cli/node_modules/@prisma/engines/migration-engine-debian-openssl-1.1.x) Introspection Engine : introspection-core fcbc4bb2d306c86c28014f596b1e8c7980af8bd4 (at ../../../../../home/kranga/.nvm/versions/node/v15.5.0/lib/node_modules/@prisma/cli/node_modules/@prisma/engines/introspection-engine-debian-openssl-1.1.x) Format Binary : prisma-fmt fcbc4bb2d306c86c28014f596b1e8c7980af8bd4 (at ../../../../../home/kranga/.nvm/versions/node/v15.5.0/lib/node_modules/@prisma/cli/node_modules/@prisma/engines/prisma-fmt-debian-openssl-1.1.x) Studio : 0.329.0 ```
1.0
A simple many to many relationship causes PANIC: 1 - ## Bug description Trying to implement a simple many to many relationship results in findMany causing panic ## How to reproduce I used the following schema to create a database: ```prisma datasource db { provider = "postgresql" url = "postgresql://[snipped]" } generator client { provider = "prisma-client-js" } model Image { hash String @id tags Tag[] @relation(name: "tags") } model Tag { name String @id images Image[] @relation(name: "tags") } ``` Then lanched prisma studio and created two Image instances then saved them and created a single tag, then I associated both images with that tag. After that any prisma.image.findMany query that includes or selects images fails with a panic message (including those of prisma studio). ## Expected behavior The query should not fail since it is almost a copy of a documentation example. ```prisma //Copied from https://www.prisma.io/docs/support/help-articles/working-with-many-to-many-relations model Post { id Int @id @default(autoincrement()) title String tags Tag[] } model Tag { id Int @id @default(autoincrement()) name String @unique posts Post[] } ``` ## Prisma information Included above ## Environment & setup <!-- In which environment does the problem occur --> - OS: Linux (Manjaro) - Database: PostgreSQL - Node.js version: 15.5.0 - Prisma version: <!--[Run `prisma -v` to see your Prisma version and paste it between the ´´´]--> ``` @prisma/cli : 2.13.1 @prisma/client : 2.13.1 Current platform : debian-openssl-1.1.x Query Engine : query-engine fcbc4bb2d306c86c28014f596b1e8c7980af8bd4 (at ../../../../../home/kranga/.nvm/versions/node/v15.5.0/lib/node_modules/@prisma/cli/node_modules/@prisma/engines/query-engine-debian-openssl-1.1.x) Migration Engine : migration-engine-cli fcbc4bb2d306c86c28014f596b1e8c7980af8bd4 (at ../../../../../home/kranga/.nvm/versions/node/v15.5.0/lib/node_modules/@prisma/cli/node_modules/@prisma/engines/migration-engine-debian-openssl-1.1.x) Introspection Engine : introspection-core fcbc4bb2d306c86c28014f596b1e8c7980af8bd4 (at ../../../../../home/kranga/.nvm/versions/node/v15.5.0/lib/node_modules/@prisma/cli/node_modules/@prisma/engines/introspection-engine-debian-openssl-1.1.x) Format Binary : prisma-fmt fcbc4bb2d306c86c28014f596b1e8c7980af8bd4 (at ../../../../../home/kranga/.nvm/versions/node/v15.5.0/lib/node_modules/@prisma/cli/node_modules/@prisma/engines/prisma-fmt-debian-openssl-1.1.x) Studio : 0.329.0 ```
process
a simple many to many relationship causes panic bug description trying to implement a simple many to many relationship results in findmany causing panic how to reproduce i used the following schema to create a database prisma datasource db provider postgresql url postgresql generator client provider prisma client js model image hash string id tags tag relation name tags model tag name string id images image relation name tags then lanched prisma studio and created two image instances then saved them and created a single tag then i associated both images with that tag after that any prisma image findmany query that includes or selects images fails with a panic message including those of prisma studio expected behavior the query should not fail since it is almost a copy of a documentation example prisma copied from model post id int id default autoincrement title string tags tag model tag id int id default autoincrement name string unique posts post prisma information included above environment setup os linux manjaro database postgresql node js version prisma version prisma cli prisma client current platform debian openssl x query engine query engine at home kranga nvm versions node lib node modules prisma cli node modules prisma engines query engine debian openssl x migration engine migration engine cli at home kranga nvm versions node lib node modules prisma cli node modules prisma engines migration engine debian openssl x introspection engine introspection core at home kranga nvm versions node lib node modules prisma cli node modules prisma engines introspection engine debian openssl x format binary prisma fmt at home kranga nvm versions node lib node modules prisma cli node modules prisma engines prisma fmt debian openssl x studio
1
16,926
22,272,992,774
IssuesEvent
2022-06-10 14:01:38
hashgraph/hedera-json-rpc-relay
https://api.github.com/repos/hashgraph/hedera-json-rpc-relay
closed
Startup logs should inform correct setup
enhancement P1 process
### Problem Currently the initial set of logs don't always make it clear what configurations have been set and whether correct operation of the relay is expected. ### Solution Improve logs to clearly show - mIrror node url - consensus node endpoints ### Alternatives _No response_
1.0
Startup logs should inform correct setup - ### Problem Currently the initial set of logs don't always make it clear what configurations have been set and whether correct operation of the relay is expected. ### Solution Improve logs to clearly show - mIrror node url - consensus node endpoints ### Alternatives _No response_
process
startup logs should inform correct setup problem currently the initial set of logs don t always make it clear what configurations have been set and whether correct operation of the relay is expected solution improve logs to clearly show mirror node url consensus node endpoints alternatives no response
1
20,548
27,204,387,404
IssuesEvent
2023-02-20 12:01:37
GIScience/sketch-map-tool
https://api.github.com/repos/GIScience/sketch-map-tool
closed
Sketch Maps in portrait orientation are not correctly georeferenced
bug component:upload-processing
Both GeoTIFF and vector data are rotated by 90 degrees
1.0
Sketch Maps in portrait orientation are not correctly georeferenced - Both GeoTIFF and vector data are rotated by 90 degrees
process
sketch maps in portrait orientation are not correctly georeferenced both geotiff and vector data are rotated by degrees
1
12,589
14,991,895,956
IssuesEvent
2021-01-29 09:05:04
panther-labs/panther
https://api.github.com/repos/panther-labs/panther
opened
Define pattern/framework for Lambdas to create System Health alarms
p1 story team:data processing
### Description Define framework for Lambdas to create System Health alarms. The same pattern/framework will be used in all parts of the system that require to set up their System Health alarms. ### Related Services All ### Designs Not needed ### Acceptance Criteria - Implementation of a metrics framework that our system can use to publish metrics to CloudWatch - The framework should make use of CloudWatch alarms - The framework is used in one part of the system to demonstrate its usage
1.0
Define pattern/framework for Lambdas to create System Health alarms - ### Description Define framework for Lambdas to create System Health alarms. The same pattern/framework will be used in all parts of the system that require to set up their System Health alarms. ### Related Services All ### Designs Not needed ### Acceptance Criteria - Implementation of a metrics framework that our system can use to publish metrics to CloudWatch - The framework should make use of CloudWatch alarms - The framework is used in one part of the system to demonstrate its usage
process
define pattern framework for lambdas to create system health alarms description define framework for lambdas to create system health alarms the same pattern framework will be used in all parts of the system that require to set up their system health alarms related services all designs not needed acceptance criteria implementation of a metrics framework that our system can use to publish metrics to cloudwatch the framework should make use of cloudwatch alarms the framework is used in one part of the system to demonstrate its usage
1
18,181
24,233,374,486
IssuesEvent
2022-09-26 20:24:37
GoogleCloudPlatform/cloud-ops-sandbox
https://api.github.com/repos/GoogleCloudPlatform/cloud-ops-sandbox
closed
Dependency Dashboard
priority: p2 type: process
This issue lists Renovate updates and detected dependencies. Read the [Dependency Dashboard](https://docs.renovatebot.com/key-concepts/dashboard/) docs to learn more. ## Repository problems These problems occurred while renovating this repository. - WARN: Base branch does not exist - skipping This repository currently has no open or pending branches. ## Detected dependencies --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
1.0
Dependency Dashboard - This issue lists Renovate updates and detected dependencies. Read the [Dependency Dashboard](https://docs.renovatebot.com/key-concepts/dashboard/) docs to learn more. ## Repository problems These problems occurred while renovating this repository. - WARN: Base branch does not exist - skipping This repository currently has no open or pending branches. ## Detected dependencies --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
process
dependency dashboard this issue lists renovate updates and detected dependencies read the docs to learn more repository problems these problems occurred while renovating this repository warn base branch does not exist skipping this repository currently has no open or pending branches detected dependencies check this box to trigger a request for renovate to run again on this repository
1
534,315
15,614,194,124
IssuesEvent
2021-03-19 17:25:50
canonical-web-and-design/ubuntu.com
https://api.github.com/repos/canonical-web-and-design/ubuntu.com
closed
Take down /16-04/gcp page
Priority: High
Can you take this page down? It's not accurate and we still need to update and confirm the page copy with various teams. --- *Reported from: https://ubuntu.com/16-04/gcp*
1.0
Take down /16-04/gcp page - Can you take this page down? It's not accurate and we still need to update and confirm the page copy with various teams. --- *Reported from: https://ubuntu.com/16-04/gcp*
non_process
take down gcp page can you take this page down it s not accurate and we still need to update and confirm the page copy with various teams reported from
0
22,403
31,142,291,001
IssuesEvent
2023-08-16 01:44:39
cypress-io/cypress
https://api.github.com/repos/cypress-io/cypress
closed
Flaky test: AssertionError: Timed out retrying after 10000ms: Expected to find content: 'Spec not found' but never did.
OS: linux process: flaky test topic: flake ❄️ stage: flake stale
### Link to dashboard or CircleCI failure https://app.circleci.com/pipelines/github/cypress-io/cypress/41301/workflows/ed35f5b9-63a5-409c-8893-f0cd8a5bf952/jobs/1709537 ### Link to failing test in GitHub https://github.com/cypress-io/cypress/blob/develop/packages/app/cypress/e2e/cypress-in-cypress-component.cy.ts#L81 ### Analysis <img width="1131" alt="Screen Shot 2022-08-05 at 12 43 53 PM" src="https://user-images.githubusercontent.com/26726429/183149599-341591f6-c557-4ea4-929f-b11e4e40ff43.png"> ### Cypress Version 10.4.0 ### Other Search for this issue number in the codebase to find the test(s) skipped until this issue is fixed
1.0
Flaky test: AssertionError: Timed out retrying after 10000ms: Expected to find content: 'Spec not found' but never did. - ### Link to dashboard or CircleCI failure https://app.circleci.com/pipelines/github/cypress-io/cypress/41301/workflows/ed35f5b9-63a5-409c-8893-f0cd8a5bf952/jobs/1709537 ### Link to failing test in GitHub https://github.com/cypress-io/cypress/blob/develop/packages/app/cypress/e2e/cypress-in-cypress-component.cy.ts#L81 ### Analysis <img width="1131" alt="Screen Shot 2022-08-05 at 12 43 53 PM" src="https://user-images.githubusercontent.com/26726429/183149599-341591f6-c557-4ea4-929f-b11e4e40ff43.png"> ### Cypress Version 10.4.0 ### Other Search for this issue number in the codebase to find the test(s) skipped until this issue is fixed
process
flaky test assertionerror timed out retrying after expected to find content spec not found but never did link to dashboard or circleci failure link to failing test in github analysis img width alt screen shot at pm src cypress version other search for this issue number in the codebase to find the test s skipped until this issue is fixed
1
8,477
11,643,051,604
IssuesEvent
2020-02-29 11:05:11
tikv/tikv
https://api.github.com/repos/tikv/tikv
opened
UCP: Migrate scalar function `SubstringIndex` from TiDB
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
## Description Port the scalar function `SubstringIndex` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @breeswish ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
2.0
UCP: Migrate scalar function `SubstringIndex` from TiDB - ## Description Port the scalar function `SubstringIndex` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @breeswish ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
process
ucp migrate scalar function substringindex from tidb description port the scalar function substringindex from tidb to coprocessor score mentor s breeswish recommended skills rust programming learning materials already implemented expressions ported from tidb
1
6,735
9,866,780,500
IssuesEvent
2019-06-21 08:31:32
googleapis/google-cloud-python
https://api.github.com/repos/googleapis/google-cloud-python
closed
Missing system tests after pubsub redesign
api: pubsub testing triaged for GA type: process
The redesign effort (#3859) left behind only very minimal system tests. The older implementation had much broader system test coverage: - Listing topics and subscriptions in the client's project. - Creating subscriptions with non-default settings. - Listing subscriptions bound to a topic. - Setting / getting IAM policy for topics and subscriptions - Creating / seeking snapshots. I noticed the missing tests today when I went looking for examples of updating a topic's IAM policy.
1.0
Missing system tests after pubsub redesign - The redesign effort (#3859) left behind only very minimal system tests. The older implementation had much broader system test coverage: - Listing topics and subscriptions in the client's project. - Creating subscriptions with non-default settings. - Listing subscriptions bound to a topic. - Setting / getting IAM policy for topics and subscriptions - Creating / seeking snapshots. I noticed the missing tests today when I went looking for examples of updating a topic's IAM policy.
process
missing system tests after pubsub redesign the redesign effort left behind only very minimal system tests the older implementation had much broader system test coverage listing topics and subscriptions in the client s project creating subscriptions with non default settings listing subscriptions bound to a topic setting getting iam policy for topics and subscriptions creating seeking snapshots i noticed the missing tests today when i went looking for examples of updating a topic s iam policy
1
99,989
4,075,204,750
IssuesEvent
2016-05-29 01:39:16
revel/revel
https://api.github.com/repos/revel/revel
closed
Update references to github.com/revel/config
priority-should topic-config type-enhancement
As mentioned in the documentation, update the reference in the code base as well. - [x] First synchronize the fork from parent `github.com/robfig/config` - [x] Update `github.com/robfig/config` => `github.com/revel/config`
1.0
Update references to github.com/revel/config - As mentioned in the documentation, update the reference in the code base as well. - [x] First synchronize the fork from parent `github.com/robfig/config` - [x] Update `github.com/robfig/config` => `github.com/revel/config`
non_process
update references to github com revel config as mentioned in the documentation update the reference in the code base as well first synchronize the fork from parent github com robfig config update github com robfig config github com revel config
0
6,043
7,469,519,908
IssuesEvent
2018-04-02 23:18:41
Microsoft/vscode-cpptools
https://api.github.com/repos/Microsoft/vscode-cpptools
closed
Please add support for intellisence for /clr
Feature Request Language Service fixed (release pending) quick fix
I have a C++ project with code that must be compiled with /clr. Please add support for intellisence for /clr
1.0
Please add support for intellisence for /clr - I have a C++ project with code that must be compiled with /clr. Please add support for intellisence for /clr
non_process
please add support for intellisence for clr i have a c project with code that must be compiled with clr please add support for intellisence for clr
0
15,300
19,325,017,210
IssuesEvent
2021-12-14 10:22:33
decidim/decidim
https://api.github.com/repos/decidim/decidim
opened
Removing a scope from selected scope picker doesn't work
type: bug module: participatory processes
**Describe the bug** When I select a scope from the scope picker, for instance from participatory proceses, it doesn't behave correctly when refreshing **To Reproduce** Steps to reproduce the behavior: 1. Go to /processes 2. Click on "Select a scope" 3. Click on any scope 4. Error 1: if the page has scroll it goes to top 5. Refresh the page 6. Error 2: see that the label has lost the CSS design 7. Click on the label to remove. 8. Error 3: See that it disappears, but the filter selection doesn't change 8. Refresh the page 9. Error 4: see that the label/scope is there again **Expected behavior** I expect 3 things: 1. When I click in a scope in this picker and there's scroll, it shouldn't go to the top of the page (Error 1) 2. When I refresh after selecting a scope, I shouldn't lose the label's CSS (Error 2) 3. When I remove a filtered scope, it should show the change (Error 3) 4. When I refresh after removing a filtered scope, it should be removed for good (Error 4) **Screenshots** ![scope-picker-bug](https://user-images.githubusercontent.com/717367/145979763-88ac2b36-281f-46aa-b033-694e917378ef.gif) **Extra data (please complete the following information):** - Device: Desktop - Browser: Firefox - Decidim Version: 0.26.0.dev - Decidim installation: Codegram staging
1.0
Removing a scope from selected scope picker doesn't work - **Describe the bug** When I select a scope from the scope picker, for instance from participatory proceses, it doesn't behave correctly when refreshing **To Reproduce** Steps to reproduce the behavior: 1. Go to /processes 2. Click on "Select a scope" 3. Click on any scope 4. Error 1: if the page has scroll it goes to top 5. Refresh the page 6. Error 2: see that the label has lost the CSS design 7. Click on the label to remove. 8. Error 3: See that it disappears, but the filter selection doesn't change 8. Refresh the page 9. Error 4: see that the label/scope is there again **Expected behavior** I expect 3 things: 1. When I click in a scope in this picker and there's scroll, it shouldn't go to the top of the page (Error 1) 2. When I refresh after selecting a scope, I shouldn't lose the label's CSS (Error 2) 3. When I remove a filtered scope, it should show the change (Error 3) 4. When I refresh after removing a filtered scope, it should be removed for good (Error 4) **Screenshots** ![scope-picker-bug](https://user-images.githubusercontent.com/717367/145979763-88ac2b36-281f-46aa-b033-694e917378ef.gif) **Extra data (please complete the following information):** - Device: Desktop - Browser: Firefox - Decidim Version: 0.26.0.dev - Decidim installation: Codegram staging
process
removing a scope from selected scope picker doesn t work describe the bug when i select a scope from the scope picker for instance from participatory proceses it doesn t behave correctly when refreshing to reproduce steps to reproduce the behavior go to processes click on select a scope click on any scope error if the page has scroll it goes to top refresh the page error see that the label has lost the css design click on the label to remove error see that it disappears but the filter selection doesn t change refresh the page error see that the label scope is there again expected behavior i expect things when i click in a scope in this picker and there s scroll it shouldn t go to the top of the page error when i refresh after selecting a scope i shouldn t lose the label s css error when i remove a filtered scope it should show the change error when i refresh after removing a filtered scope it should be removed for good error screenshots extra data please complete the following information device desktop browser firefox decidim version dev decidim installation codegram staging
1
11,858
14,665,036,366
IssuesEvent
2020-12-29 13:23:39
modi-w/AutoVersionsDB
https://api.github.com/repos/modi-w/AutoVersionsDB
opened
Handle State Data Files
area-Core area-Tests area-UI process-discussion type-enhancement
**The Problem** The main purpose of this tool is to automate the process of setting the database in sync with the specific location (specific commit) with code on the source control. In other words: make it easy to sync the DB state with the code state no matter which branch\commit the code. But sometimes in some projects, the database is not the only data state of the system. Sometimes we have external files that define the data state of the system. For example: for accounting modules in the system, we have some pdf files for invoices. **Solution** This section should be for discussion. One option maybe adds kind of "script folders" for: 1. DevDummyData state data files 2. Repeatable state data files **Action Items:** 1. 2. 3. **Updates** 1.
1.0
Handle State Data Files - **The Problem** The main purpose of this tool is to automate the process of setting the database in sync with the specific location (specific commit) with code on the source control. In other words: make it easy to sync the DB state with the code state no matter which branch\commit the code. But sometimes in some projects, the database is not the only data state of the system. Sometimes we have external files that define the data state of the system. For example: for accounting modules in the system, we have some pdf files for invoices. **Solution** This section should be for discussion. One option maybe adds kind of "script folders" for: 1. DevDummyData state data files 2. Repeatable state data files **Action Items:** 1. 2. 3. **Updates** 1.
process
handle state data files the problem the main purpose of this tool is to automate the process of setting the database in sync with the specific location specific commit with code on the source control in other words make it easy to sync the db state with the code state no matter which branch commit the code but sometimes in some projects the database is not the only data state of the system sometimes we have external files that define the data state of the system for example for accounting modules in the system we have some pdf files for invoices solution this section should be for discussion one option maybe adds kind of script folders for devdummydata state data files repeatable state data files action items updates
1
20,915
27,754,011,586
IssuesEvent
2023-03-15 23:54:40
dDevTech/tapas-top-frontend
https://api.github.com/repos/dDevTech/tapas-top-frontend
closed
Modificación pagina de registro 20/03/2023
pending in process require testing
Como jhipster nos ha creado las paginas por defecto, se debe modificar la de registro para que el botón Crear Cuenta sea continuar y nos lleva a la pagina register-account-info. Debemos haber guardado la información de registro con redux de forma temporal en vez de enviarla directamente a la base de datos
1.0
Modificación pagina de registro 20/03/2023 - Como jhipster nos ha creado las paginas por defecto, se debe modificar la de registro para que el botón Crear Cuenta sea continuar y nos lleva a la pagina register-account-info. Debemos haber guardado la información de registro con redux de forma temporal en vez de enviarla directamente a la base de datos
process
modificación pagina de registro como jhipster nos ha creado las paginas por defecto se debe modificar la de registro para que el botón crear cuenta sea continuar y nos lleva a la pagina register account info debemos haber guardado la información de registro con redux de forma temporal en vez de enviarla directamente a la base de datos
1
18,765
24,669,222,636
IssuesEvent
2022-10-18 12:40:31
MasterPlayer/adxl345-sv
https://api.github.com/repos/MasterPlayer/adxl345-sv
opened
FIFO interrupt processing support
enhancement hardware process software process
There are needed for special processing for FIFO interrupts, because sw must read series of samples. Possible mechanism may be realized as next scheme:
2.0
FIFO interrupt processing support - There are needed for special processing for FIFO interrupts, because sw must read series of samples. Possible mechanism may be realized as next scheme:
process
fifo interrupt processing support there are needed for special processing for fifo interrupts because sw must read series of samples possible mechanism may be realized as next scheme
1
21,077
28,019,962,159
IssuesEvent
2023-03-28 04:03:29
0xPolygonMiden/miden-vm
https://api.github.com/repos/0xPolygonMiden/miden-vm
closed
Replace MerkleSets in the advice provider with MerkleStore
processor
Now that we have [MerkleStore](https://github.com/0xPolygonMiden/crypto/blob/next/src/merkle/store.rs) implemented in `miden-crypto`, we should use it for instead of a map of `MerkleSet`'s in [MemAdviceProvider](https://github.com/0xPolygonMiden/miden-vm/blob/main/processor/src/advice/mem_provider.rs). We should also probable rename the fields like this: * `tape` -> `stack` * `values` -> `map` * `sets` -> `store` So, the `MemAdviceProvider` struct could look like this: ```Rust pub struct MemAdviceProvider { step: u32, stack: Vec<Felt>, map: BTreeMap<[u8; 32], Vec<Felt>>, store: MerkleStore, } ``` We might also want to consider renaming `tape` into `stack` in the [AdviceProvider](https://github.com/0xPolygonMiden/miden-vm/blob/main/processor/src/advice/mod.rs#L52) trait as well.
1.0
Replace MerkleSets in the advice provider with MerkleStore - Now that we have [MerkleStore](https://github.com/0xPolygonMiden/crypto/blob/next/src/merkle/store.rs) implemented in `miden-crypto`, we should use it for instead of a map of `MerkleSet`'s in [MemAdviceProvider](https://github.com/0xPolygonMiden/miden-vm/blob/main/processor/src/advice/mem_provider.rs). We should also probable rename the fields like this: * `tape` -> `stack` * `values` -> `map` * `sets` -> `store` So, the `MemAdviceProvider` struct could look like this: ```Rust pub struct MemAdviceProvider { step: u32, stack: Vec<Felt>, map: BTreeMap<[u8; 32], Vec<Felt>>, store: MerkleStore, } ``` We might also want to consider renaming `tape` into `stack` in the [AdviceProvider](https://github.com/0xPolygonMiden/miden-vm/blob/main/processor/src/advice/mod.rs#L52) trait as well.
process
replace merklesets in the advice provider with merklestore now that we have implemented in miden crypto we should use it for instead of a map of merkleset s in we should also probable rename the fields like this tape stack values map sets store so the memadviceprovider struct could look like this rust pub struct memadviceprovider step stack vec map btreemap store merklestore we might also want to consider renaming tape into stack in the trait as well
1
327,932
24,162,319,206
IssuesEvent
2022-09-22 12:39:32
giantswarm/roadmap
https://api.github.com/repos/giantswarm/roadmap
closed
Deprecate CRDs
topic/documentation team/rainbow topic/crd
With CAPI, many CRDs we currently document in [our docs](https://docs.giantswarm.io/ui-api/management-api/crd/) are no longer used. I'd like to mark these CRDs as deprecated. This way it's easier to understand for everyone which CRDs have a future.
1.0
Deprecate CRDs - With CAPI, many CRDs we currently document in [our docs](https://docs.giantswarm.io/ui-api/management-api/crd/) are no longer used. I'd like to mark these CRDs as deprecated. This way it's easier to understand for everyone which CRDs have a future.
non_process
deprecate crds with capi many crds we currently document in are no longer used i d like to mark these crds as deprecated this way it s easier to understand for everyone which crds have a future
0
15,308
19,400,850,809
IssuesEvent
2021-12-19 06:13:31
ethereum/EIPs
https://api.github.com/repos/ethereum/EIPs
closed
Add mission statement
type: Meta type: EIP1 (Process) stale
Presently the ethereum/EIPs project does not have a mission statement. --- <strike>Recently something changed and now the majority of EIPs here have no path to become "final" standards. Pull request #1100 addresses that issue.</strike> However, one of the EIP editors (the people with commit access here) mentioned that #1100 is not urgent. There are no remaining complaints on #1100, it has EIP editor endorsements, but it is not merged. I reviewed the project README.md and was hoping to find something like "our goal is to discuss and pass high-quality standards reflecting established best practices in the community." So I could tell this person that #1100 is urgent (because presently, standards are prevented from passing). Alas no such line exists, in fact, there is nothing in the README.md that explains why we are contributing here. **It is much easier to set expectations for each other in this project if we have a clearly defined goal. And we should state that goal in the README.md.**
1.0
Add mission statement - Presently the ethereum/EIPs project does not have a mission statement. --- <strike>Recently something changed and now the majority of EIPs here have no path to become "final" standards. Pull request #1100 addresses that issue.</strike> However, one of the EIP editors (the people with commit access here) mentioned that #1100 is not urgent. There are no remaining complaints on #1100, it has EIP editor endorsements, but it is not merged. I reviewed the project README.md and was hoping to find something like "our goal is to discuss and pass high-quality standards reflecting established best practices in the community." So I could tell this person that #1100 is urgent (because presently, standards are prevented from passing). Alas no such line exists, in fact, there is nothing in the README.md that explains why we are contributing here. **It is much easier to set expectations for each other in this project if we have a clearly defined goal. And we should state that goal in the README.md.**
process
add mission statement presently the ethereum eips project does not have a mission statement recently something changed and now the majority of eips here have no path to become final standards pull request addresses that issue however one of the eip editors the people with commit access here mentioned that is not urgent there are no remaining complaints on it has eip editor endorsements but it is not merged i reviewed the project readme md and was hoping to find something like our goal is to discuss and pass high quality standards reflecting established best practices in the community so i could tell this person that is urgent because presently standards are prevented from passing alas no such line exists in fact there is nothing in the readme md that explains why we are contributing here it is much easier to set expectations for each other in this project if we have a clearly defined goal and we should state that goal in the readme md
1
11,617
14,480,903,733
IssuesEvent
2020-12-10 11:50:57
Arch666Angel/mods
https://api.github.com/repos/Arch666Angel/mods
closed
[BUG] Agriculture Modules can be put in labs
Angels Bio Processing Impact: Bug
Agriculture Modules can be put in labs. Tested with just Angel's Mods (no overhauls) as well as with Bob's Technology. Issue applies to all labs. With bobs: lab, lab 2, alien lab, modules lab.
1.0
[BUG] Agriculture Modules can be put in labs - Agriculture Modules can be put in labs. Tested with just Angel's Mods (no overhauls) as well as with Bob's Technology. Issue applies to all labs. With bobs: lab, lab 2, alien lab, modules lab.
process
agriculture modules can be put in labs agriculture modules can be put in labs tested with just angel s mods no overhauls as well as with bob s technology issue applies to all labs with bobs lab lab alien lab modules lab
1
342,565
30,627,464,733
IssuesEvent
2023-07-24 12:31:58
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
opened
Fix linalg.test_tensorflow_matrix_transpose
TensorFlow Frontend Sub Task Failing Test
| | | |---|---| |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5634736017/job/15264943658"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5634736017/job/15264943658"><img src=https://img.shields.io/badge/-success-success></a> |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5634736017/job/15264943658"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5634736017/job/15264943658"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5634736017/job/15264943658"><img src=https://img.shields.io/badge/-failure-red></a>
1.0
Fix linalg.test_tensorflow_matrix_transpose - | | | |---|---| |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5634736017/job/15264943658"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5634736017/job/15264943658"><img src=https://img.shields.io/badge/-success-success></a> |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5634736017/job/15264943658"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5634736017/job/15264943658"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5634736017/job/15264943658"><img src=https://img.shields.io/badge/-failure-red></a>
non_process
fix linalg test tensorflow matrix transpose jax a href src numpy a href src tensorflow a href src torch a href src paddle a href src
0
21,934
30,446,677,667
IssuesEvent
2023-07-15 19:08:58
h4sh5/pypi-auto-scanner
https://api.github.com/repos/h4sh5/pypi-auto-scanner
opened
pyutils 0.0.1b5 has 2 GuardDog issues
guarddog typosquatting silent-process-execution
https://pypi.org/project/pyutils https://inspector.pypi.io/project/pyutils ```{ "dependency": "pyutils", "version": "0.0.1b5", "result": { "issues": 2, "errors": {}, "results": { "typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: python-utils, pytils", "silent-process-execution": [ { "location": "pyutils-0.0.1b5/src/pyutils/exec_utils.py:204", "code": " subproc = subprocess.Popen(\n args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n )", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmpx5rt84g_/pyutils" } }```
1.0
pyutils 0.0.1b5 has 2 GuardDog issues - https://pypi.org/project/pyutils https://inspector.pypi.io/project/pyutils ```{ "dependency": "pyutils", "version": "0.0.1b5", "result": { "issues": 2, "errors": {}, "results": { "typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: python-utils, pytils", "silent-process-execution": [ { "location": "pyutils-0.0.1b5/src/pyutils/exec_utils.py:204", "code": " subproc = subprocess.Popen(\n args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n )", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmpx5rt84g_/pyutils" } }```
process
pyutils has guarddog issues dependency pyutils version result issues errors results typosquatting this package closely ressembles the following package names and might be a typosquatting attempt python utils pytils silent process execution location pyutils src pyutils exec utils py code subproc subprocess popen n args n stdin subprocess devnull n stdout subprocess devnull n stderr subprocess devnull n message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp pyutils
1
3,901
6,822,593,502
IssuesEvent
2017-11-07 20:38:05
nodejs/node
https://api.github.com/repos/nodejs/node
closed
Improve ChildProcess::killed property behaviour. Or update documentation.
child_process
<!-- Thank you for reporting an issue. This issue tracker is for bugs and issues found within Node.js core. If you require more general support please file an issue on our help repo. https://github.com/nodejs/help Please fill in as much of the template below as you're able. Subsystem: doc If possible, please provide code that demonstrates the problem, keeping it as simple and free of external dependencies as you are able. --> * **Subsystem**: doc and child_process <!-- Enter your issue details below this comment. --> According to documentation (https://nodejs.org/dist/latest-v9.x/docs/api/child_process.html#child_process_subprocess_killed) the `killed` property has invalid semantic in comparison to ChildProcess::kill method behaviour (internal/child_process.js) Concretely, the problematic code is ``` var err = this._handle.kill(signal); if (err === 0) { /* Success. */ this.killed = true; return true; } ``` After the `this._handle.kill(signal);` method call we have no idea whether the 3-d party process was killed, or it just received a signal, processed it and continues working (it maybe be possible even with such 'kill' signals like SIGINT and SIGTERM). So I propose to discuss what can be done in this direction to improve expected behaviour. The easiest solution (and maybe the right one) is to move `this.killed = true;` line into `_handle.onexit` function and fix to the following code: ``` if (signalCode) { this.signalCode = signalCode; if (this._killWasCalled) { (1) this.killed = true; } } else { this.exitCode = exitCode; } ``` The if block (1) is needed cuz now documentation says '...indicates whether the child process was successfully terminated using `subprocess.kill()`'. We can eliminate this block, but we will need to update documentation also. Strictly speaking, the `killed` field isn't very helpful in terms of identifying when process will exit. We usually use `exit`event for that. But, nevertheless, it's strange to see childProcess.killed === true, when in reality process works fine.
1.0
Improve ChildProcess::killed property behaviour. Or update documentation. - <!-- Thank you for reporting an issue. This issue tracker is for bugs and issues found within Node.js core. If you require more general support please file an issue on our help repo. https://github.com/nodejs/help Please fill in as much of the template below as you're able. Subsystem: doc If possible, please provide code that demonstrates the problem, keeping it as simple and free of external dependencies as you are able. --> * **Subsystem**: doc and child_process <!-- Enter your issue details below this comment. --> According to documentation (https://nodejs.org/dist/latest-v9.x/docs/api/child_process.html#child_process_subprocess_killed) the `killed` property has invalid semantic in comparison to ChildProcess::kill method behaviour (internal/child_process.js) Concretely, the problematic code is ``` var err = this._handle.kill(signal); if (err === 0) { /* Success. */ this.killed = true; return true; } ``` After the `this._handle.kill(signal);` method call we have no idea whether the 3-d party process was killed, or it just received a signal, processed it and continues working (it maybe be possible even with such 'kill' signals like SIGINT and SIGTERM). So I propose to discuss what can be done in this direction to improve expected behaviour. The easiest solution (and maybe the right one) is to move `this.killed = true;` line into `_handle.onexit` function and fix to the following code: ``` if (signalCode) { this.signalCode = signalCode; if (this._killWasCalled) { (1) this.killed = true; } } else { this.exitCode = exitCode; } ``` The if block (1) is needed cuz now documentation says '...indicates whether the child process was successfully terminated using `subprocess.kill()`'. We can eliminate this block, but we will need to update documentation also. Strictly speaking, the `killed` field isn't very helpful in terms of identifying when process will exit. We usually use `exit`event for that. But, nevertheless, it's strange to see childProcess.killed === true, when in reality process works fine.
process
improve childprocess killed property behaviour or update documentation thank you for reporting an issue this issue tracker is for bugs and issues found within node js core if you require more general support please file an issue on our help repo please fill in as much of the template below as you re able subsystem doc if possible please provide code that demonstrates the problem keeping it as simple and free of external dependencies as you are able subsystem doc and child process according to documentation the killed property has invalid semantic in comparison to childprocess kill method behaviour internal child process js concretely the problematic code is var err this handle kill signal if err success this killed true return true after the this handle kill signal method call we have no idea whether the d party process was killed or it just received a signal processed it and continues working it maybe be possible even with such kill signals like sigint and sigterm so i propose to discuss what can be done in this direction to improve expected behaviour the easiest solution and maybe the right one is to move this killed true line into handle onexit function and fix to the following code if signalcode this signalcode signalcode if this killwascalled this killed true else this exitcode exitcode the if block is needed cuz now documentation says indicates whether the child process was successfully terminated using subprocess kill we can eliminate this block but we will need to update documentation also strictly speaking the killed field isn t very helpful in terms of identifying when process will exit we usually use exit event for that but nevertheless it s strange to see childprocess killed true when in reality process works fine
1
245,003
26,498,586,526
IssuesEvent
2023-01-18 08:30:05
dedis/d-voting
https://api.github.com/repos/dedis/d-voting
closed
THREAT - A user who is not an admin or operator cannot vote.
security issue web backend
## Scenario Every action from a user who is not an admin or operator will become unauthorized, including casting a vote. An authenticated user should able to cast a vote even if they are not an operator or admin ## Source web/backend/src/Server.ts ```js // Secure /api/evoting to admins and operators app.use('/api/evoting/*', (req, res, next) => { if (!isAuthorized(req.session.userid, SUBJECT_ELECTION, ACTION_CREATE)) { res.status(400).send('Unauthorized - only admins and operators allowed'); return; } next(); }); ``` ## Breaking Property Availability, Authorization ## Risk CVSS Score: [5.7/10](https://www.first.org/cvss/calculator/3.1#CVSS:3.1/AV:N/AC:L/PR:L/UI:R/S:U/C:N/I:N/A:H) ## Mitigation Backend should let an authenticated user have the ability to vote.
True
THREAT - A user who is not an admin or operator cannot vote. - ## Scenario Every action from a user who is not an admin or operator will become unauthorized, including casting a vote. An authenticated user should able to cast a vote even if they are not an operator or admin ## Source web/backend/src/Server.ts ```js // Secure /api/evoting to admins and operators app.use('/api/evoting/*', (req, res, next) => { if (!isAuthorized(req.session.userid, SUBJECT_ELECTION, ACTION_CREATE)) { res.status(400).send('Unauthorized - only admins and operators allowed'); return; } next(); }); ``` ## Breaking Property Availability, Authorization ## Risk CVSS Score: [5.7/10](https://www.first.org/cvss/calculator/3.1#CVSS:3.1/AV:N/AC:L/PR:L/UI:R/S:U/C:N/I:N/A:H) ## Mitigation Backend should let an authenticated user have the ability to vote.
non_process
threat a user who is not an admin or operator cannot vote scenario every action from a user who is not an admin or operator will become unauthorized including casting a vote an authenticated user should able to cast a vote even if they are not an operator or admin source web backend src server ts js secure api evoting to admins and operators app use api evoting req res next if isauthorized req session userid subject election action create res status send unauthorized only admins and operators allowed return next breaking property availability authorization risk cvss score mitigation backend should let an authenticated user have the ability to vote
0
11,251
14,018,637,091
IssuesEvent
2020-10-29 17:05:15
fluent/fluent-bit
https://api.github.com/repos/fluent/fluent-bit
closed
parser: support subsecond resolution with colon (%s:%L)
work-in-process
## Bug Report **Describe the bug** I have a 3rd party application (Forgerock OpenAM) that writes timestamps as follows in some log files: ``` amAuthInternal:05/29/2020 11:58:21:127 AM UTC: Thread[localhost-startStop-1,5,main]: TransactionId[3da35379-317e-41fe-b9f2-0cae565a8480-0] AuthContext::getLoginStatus() ``` Note that it uses a colon to separate subseconds, hence of the format "%S:%L". As per https://github.com/fluent/fluent-bit/issues/703, only "%S.%L" (dot) and "%S,%L" (comma) are currently supported by fluent-bit. **To Reproduce** Consider log entries as follows: ``` amAuthInternal:05/29/2020 11:58:21:127 AM UTC: Thread[localhost-startStop-1,5,main]: TransactionId[3da35379-317e-41fe-b9f2-0cae565a8480-0] AuthContext::getLoginStatus() amAuthInternal:05/29/2020 11:58:21:128 AM UTC: Thread[localhost-startStop-1,5,main]: TransactionId[3da35379-317e-41fe-b9f2-0cae565a8480-0] AuthContext::getSubject() amAuthInternal:05/29/2020 11:58:21:128 AM UTC: Thread[localhost-startStop-1,5,main]: TransactionId[3da35379-317e-41fe-b9f2-0cae565a8480-0] AuthContext::getAuthPrincipal(): [AuthPrincipal: cn=dsameuser,ou=DSAME Users,ou=am-config] ``` Parser config: ``` [PARSER] Name am-debug Format regex Regex ^(?<module>[A-Za-z0-9]*):(?<timestamp>\d{2}/\d{2}/\d{4} \d{2}:\d{2}:\d{2}:\d{3} [^\ ]* [^:]*): Thread\[(?<thread>[^\]]*)\]: TransactionId\[(?<transactionId>[^\]]*)\] (?<message>.*)$ Time_Key timestamp Time_Format %m/%d/%Y %I:%M:%S:%L %p %Z Time_Keep On ``` The result: ``` [2020/05/29 12:10:37] [ warn] [parser:am-debug] Invalid time format %m/%d/%Y %I:%M:%S:%L %p %Z for '05/29/2020 12:10:37:601 PM UTC'. ``` **Expected behavior** It should be possible to have any separator between %S and %L. This should be defined via Time_Format. **Your Environment** * Version used: 1.4.5 * Filters and plugins: tail
1.0
parser: support subsecond resolution with colon (%s:%L) - ## Bug Report **Describe the bug** I have a 3rd party application (Forgerock OpenAM) that writes timestamps as follows in some log files: ``` amAuthInternal:05/29/2020 11:58:21:127 AM UTC: Thread[localhost-startStop-1,5,main]: TransactionId[3da35379-317e-41fe-b9f2-0cae565a8480-0] AuthContext::getLoginStatus() ``` Note that it uses a colon to separate subseconds, hence of the format "%S:%L". As per https://github.com/fluent/fluent-bit/issues/703, only "%S.%L" (dot) and "%S,%L" (comma) are currently supported by fluent-bit. **To Reproduce** Consider log entries as follows: ``` amAuthInternal:05/29/2020 11:58:21:127 AM UTC: Thread[localhost-startStop-1,5,main]: TransactionId[3da35379-317e-41fe-b9f2-0cae565a8480-0] AuthContext::getLoginStatus() amAuthInternal:05/29/2020 11:58:21:128 AM UTC: Thread[localhost-startStop-1,5,main]: TransactionId[3da35379-317e-41fe-b9f2-0cae565a8480-0] AuthContext::getSubject() amAuthInternal:05/29/2020 11:58:21:128 AM UTC: Thread[localhost-startStop-1,5,main]: TransactionId[3da35379-317e-41fe-b9f2-0cae565a8480-0] AuthContext::getAuthPrincipal(): [AuthPrincipal: cn=dsameuser,ou=DSAME Users,ou=am-config] ``` Parser config: ``` [PARSER] Name am-debug Format regex Regex ^(?<module>[A-Za-z0-9]*):(?<timestamp>\d{2}/\d{2}/\d{4} \d{2}:\d{2}:\d{2}:\d{3} [^\ ]* [^:]*): Thread\[(?<thread>[^\]]*)\]: TransactionId\[(?<transactionId>[^\]]*)\] (?<message>.*)$ Time_Key timestamp Time_Format %m/%d/%Y %I:%M:%S:%L %p %Z Time_Keep On ``` The result: ``` [2020/05/29 12:10:37] [ warn] [parser:am-debug] Invalid time format %m/%d/%Y %I:%M:%S:%L %p %Z for '05/29/2020 12:10:37:601 PM UTC'. ``` **Expected behavior** It should be possible to have any separator between %S and %L. This should be defined via Time_Format. **Your Environment** * Version used: 1.4.5 * Filters and plugins: tail
process
parser support subsecond resolution with colon s l bug report describe the bug i have a party application forgerock openam that writes timestamps as follows in some log files amauthinternal am utc thread transactionid authcontext getloginstatus note that it uses a colon to separate subseconds hence of the format s l as per only s l dot and s l comma are currently supported by fluent bit to reproduce consider log entries as follows amauthinternal am utc thread transactionid authcontext getloginstatus amauthinternal am utc thread transactionid authcontext getsubject amauthinternal am utc thread transactionid authcontext getauthprincipal parser config name am debug format regex regex d d d d d d d thread transactionid time key timestamp time format m d y i m s l p z time keep on the result invalid time format m d y i m s l p z for pm utc expected behavior it should be possible to have any separator between s and l this should be defined via time format your environment version used filters and plugins tail
1
1,304
3,857,339,171
IssuesEvent
2016-04-07 05:18:49
PlagueHO/LabBuilder
https://api.github.com/repos/PlagueHO/LabBuilder
closed
Change Nano Server Package Property to expect actual package names
enhancement In Process
Currently the packages specified in the Packages property in a Nano Server VM configuration are looked up in an internal array and mapped to the actual filenames of the packages that are found on the ISO. This is not generic and will require updating every time new packages are released. This change should ensure that the actual filename of the package should be able to be specified instead.
1.0
Change Nano Server Package Property to expect actual package names - Currently the packages specified in the Packages property in a Nano Server VM configuration are looked up in an internal array and mapped to the actual filenames of the packages that are found on the ISO. This is not generic and will require updating every time new packages are released. This change should ensure that the actual filename of the package should be able to be specified instead.
process
change nano server package property to expect actual package names currently the packages specified in the packages property in a nano server vm configuration are looked up in an internal array and mapped to the actual filenames of the packages that are found on the iso this is not generic and will require updating every time new packages are released this change should ensure that the actual filename of the package should be able to be specified instead
1
25,996
12,339,624,335
IssuesEvent
2020-05-14 18:27:05
elastic/kibana
https://api.github.com/repos/elastic/kibana
opened
[Alerting] [Discuss] Modifying alert params within the executor as a migration tool
Feature:Alerting Team:Alerting Services discuss
Alerts are in beta right now, so we might be pushing a large amount of breaking changes to them. Required alert params might change between minor versions, and so alerts created in a previous version would immediately break. Can we fix this by including a migration assistant service in alert executors? For example: - Metric alert types have an interval (`timeSize` and `timeUnit`) inside each alert `criteria` (an array of alert conditions), but these are supposed to be the same for each of them. It's redundant and error-prone to repeat the same value inside each array member. ```js params: { criteria: [{ timeSize: 's' | 'm' | 'h' | 'd', timeUnit: number, ...rest }, ...additionalConditions ] } ``` - In the next minor version, we might move `timeSize` and `timeUnit` outside of the `criteria` array and into the top-level alert `params`. ```js params: { criteria: [...arrayOfConditions], timeSize, timeUnit } ``` - When the alert executor runs, it checks to see if `timeSize` and `timeUnit` are inside of the `criteria`. If they are, it will move them into the top level of params and **update the SavedObject of the alert** to reflect this change. This code can then be deprecated in the next minor version, assuming that all alerts probably migrated before then. Is something like this already possible with the available `savedObjects` service in alerts, or would a new one be needed? Is this even a good idea? I'd like some opinions.
1.0
[Alerting] [Discuss] Modifying alert params within the executor as a migration tool - Alerts are in beta right now, so we might be pushing a large amount of breaking changes to them. Required alert params might change between minor versions, and so alerts created in a previous version would immediately break. Can we fix this by including a migration assistant service in alert executors? For example: - Metric alert types have an interval (`timeSize` and `timeUnit`) inside each alert `criteria` (an array of alert conditions), but these are supposed to be the same for each of them. It's redundant and error-prone to repeat the same value inside each array member. ```js params: { criteria: [{ timeSize: 's' | 'm' | 'h' | 'd', timeUnit: number, ...rest }, ...additionalConditions ] } ``` - In the next minor version, we might move `timeSize` and `timeUnit` outside of the `criteria` array and into the top-level alert `params`. ```js params: { criteria: [...arrayOfConditions], timeSize, timeUnit } ``` - When the alert executor runs, it checks to see if `timeSize` and `timeUnit` are inside of the `criteria`. If they are, it will move them into the top level of params and **update the SavedObject of the alert** to reflect this change. This code can then be deprecated in the next minor version, assuming that all alerts probably migrated before then. Is something like this already possible with the available `savedObjects` service in alerts, or would a new one be needed? Is this even a good idea? I'd like some opinions.
non_process
modifying alert params within the executor as a migration tool alerts are in beta right now so we might be pushing a large amount of breaking changes to them required alert params might change between minor versions and so alerts created in a previous version would immediately break can we fix this by including a migration assistant service in alert executors for example metric alert types have an interval timesize and timeunit inside each alert criteria an array of alert conditions but these are supposed to be the same for each of them it s redundant and error prone to repeat the same value inside each array member js params criteria timesize s m h d timeunit number rest additionalconditions in the next minor version we might move timesize and timeunit outside of the criteria array and into the top level alert params js params criteria timesize timeunit when the alert executor runs it checks to see if timesize and timeunit are inside of the criteria if they are it will move them into the top level of params and update the savedobject of the alert to reflect this change this code can then be deprecated in the next minor version assuming that all alerts probably migrated before then is something like this already possible with the available savedobjects service in alerts or would a new one be needed is this even a good idea i d like some opinions
0
11,120
13,957,685,093
IssuesEvent
2020-10-24 08:08:31
alexanderkotsev/geoportal
https://api.github.com/repos/alexanderkotsev/geoportal
opened
PL: Unsuccessful harvesting
Geoportal Harvesting process PL - Poland
Hi, I did 2 harvesting tests, I received an e-mail confirming the start, the next day wanting to check the results and publish them, I get a new harvesting option without previous results and the possibility of publication. Regards Piotr
1.0
PL: Unsuccessful harvesting - Hi, I did 2 harvesting tests, I received an e-mail confirming the start, the next day wanting to check the results and publish them, I get a new harvesting option without previous results and the possibility of publication. Regards Piotr
process
pl unsuccessful harvesting hi i did harvesting tests i received an e mail confirming the start the next day wanting to check the results and publish them i get a new harvesting option without previous results and the possibility of publication regards piotr
1
11,682
8,467,515,872
IssuesEvent
2018-10-23 17:10:54
AOSC-Dev/aosc-os-abbs
https://api.github.com/repos/AOSC-Dev/aosc-os-abbs
closed
requests: CVE-2018-18074
security to-stable
<!-- Please remove items do not apply. --> **CVE IDs:** CVE-2018-18074 **Other security advisory IDs:** USN-3790-1 **Descriptions:** Requests could be made to expose sensitive information if it received a specially crafted HTTP header. **Patches:** https://github.com/requests/requests/commit/c45d7c49ea75133e52ab22a8e9e13173938e36ff **PoC(s):** https://github.com/requests/requests/issues/4716 **Architectural progress:** <!-- Please remove any architecture to which the security vulnerabilities do not apply. --> - [x] Data/architecture-independent (`noarch`)
True
requests: CVE-2018-18074 - <!-- Please remove items do not apply. --> **CVE IDs:** CVE-2018-18074 **Other security advisory IDs:** USN-3790-1 **Descriptions:** Requests could be made to expose sensitive information if it received a specially crafted HTTP header. **Patches:** https://github.com/requests/requests/commit/c45d7c49ea75133e52ab22a8e9e13173938e36ff **PoC(s):** https://github.com/requests/requests/issues/4716 **Architectural progress:** <!-- Please remove any architecture to which the security vulnerabilities do not apply. --> - [x] Data/architecture-independent (`noarch`)
non_process
requests cve cve ids cve other security advisory ids usn descriptions requests could be made to expose sensitive information if it received a specially crafted http header patches poc s architectural progress data architecture independent noarch
0
9,947
12,976,360,272
IssuesEvent
2020-07-21 18:38:17
googleapis/java-game-servers
https://api.github.com/repos/googleapis/java-game-servers
closed
Promote to Beta
api: gameservices status: blocked type: process
Package name: **google-cloud-gameservices** Current release: **alpha** Proposed release: **beta** ## Instructions Check the lists below, adding tests / documentation as required. Once all the "required" boxes are ticked, please create a release and close this issue. ## Required - [ ] Server API is beta or GA - [ ] Service API is public - [ ] Client surface is mostly stable (no known issues that could significantly change the surface) - [ ] All manual types and methods have comment documentation - [ ] Package name is idiomatic for the platform - [ ] At least one integration/smoke test is defined and passing - [ ] Central GitHub README lists and points to the per-API README - [ ] Per-API README links to product page on cloud.google.com - [ ] Manual code has been reviewed for API stability by repo owner ## Optional - [ ] Most common / important scenarios have descriptive samples - [ ] Public manual methods have at least one usage sample each (excluding overloads) - [ ] Per-API README includes a full description of the API - [ ] Per-API README contains at least one “getting started” sample using the most common API scenario - [ ] Manual code has been reviewed by API producer - [ ] Manual code has been reviewed by a DPE responsible for samples - [ ] 'Client LIbraries' page is added to the product documentation in 'APIs & Reference' section of the product's documentation on Cloud Site
1.0
Promote to Beta - Package name: **google-cloud-gameservices** Current release: **alpha** Proposed release: **beta** ## Instructions Check the lists below, adding tests / documentation as required. Once all the "required" boxes are ticked, please create a release and close this issue. ## Required - [ ] Server API is beta or GA - [ ] Service API is public - [ ] Client surface is mostly stable (no known issues that could significantly change the surface) - [ ] All manual types and methods have comment documentation - [ ] Package name is idiomatic for the platform - [ ] At least one integration/smoke test is defined and passing - [ ] Central GitHub README lists and points to the per-API README - [ ] Per-API README links to product page on cloud.google.com - [ ] Manual code has been reviewed for API stability by repo owner ## Optional - [ ] Most common / important scenarios have descriptive samples - [ ] Public manual methods have at least one usage sample each (excluding overloads) - [ ] Per-API README includes a full description of the API - [ ] Per-API README contains at least one “getting started” sample using the most common API scenario - [ ] Manual code has been reviewed by API producer - [ ] Manual code has been reviewed by a DPE responsible for samples - [ ] 'Client LIbraries' page is added to the product documentation in 'APIs & Reference' section of the product's documentation on Cloud Site
process
promote to beta package name google cloud gameservices current release alpha proposed release beta instructions check the lists below adding tests documentation as required once all the required boxes are ticked please create a release and close this issue required server api is beta or ga service api is public client surface is mostly stable no known issues that could significantly change the surface all manual types and methods have comment documentation package name is idiomatic for the platform at least one integration smoke test is defined and passing central github readme lists and points to the per api readme per api readme links to product page on cloud google com manual code has been reviewed for api stability by repo owner optional most common important scenarios have descriptive samples public manual methods have at least one usage sample each excluding overloads per api readme includes a full description of the api per api readme contains at least one “getting started” sample using the most common api scenario manual code has been reviewed by api producer manual code has been reviewed by a dpe responsible for samples client libraries page is added to the product documentation in apis reference section of the product s documentation on cloud site
1
14,502
17,604,346,935
IssuesEvent
2021-08-17 15:16:50
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
closed
[feature] add Create normal raster algorithm
Automatic new feature Processing Alg 3.14
Original commit: https://github.com/qgis/QGIS/commit/d6014d5dfe7d3d2754f1853c313ee16f4ee71721 by nyalldawson Unfortunately this naughty coder did not write a description... :-(
1.0
[feature] add Create normal raster algorithm - Original commit: https://github.com/qgis/QGIS/commit/d6014d5dfe7d3d2754f1853c313ee16f4ee71721 by nyalldawson Unfortunately this naughty coder did not write a description... :-(
process
add create normal raster algorithm original commit by nyalldawson unfortunately this naughty coder did not write a description
1
20,821
27,579,244,583
IssuesEvent
2023-03-08 15:08:16
ukri-excalibur/excalibur-tests
https://api.github.com/repos/ukri-excalibur/excalibur-tests
opened
Scripts for creating plots/tables of machine benchmarking
UCL postprocessing
Or "use case 3" in https://github.com/ukri-excalibur/excalibur-tests/issues/70#issue-1522882139 , duplicated here: > Several benchmark apps, run on the same machine. Again, technically the same as the first case, but might benefit from simplified script.
1.0
Scripts for creating plots/tables of machine benchmarking - Or "use case 3" in https://github.com/ukri-excalibur/excalibur-tests/issues/70#issue-1522882139 , duplicated here: > Several benchmark apps, run on the same machine. Again, technically the same as the first case, but might benefit from simplified script.
process
scripts for creating plots tables of machine benchmarking or use case in duplicated here several benchmark apps run on the same machine again technically the same as the first case but might benefit from simplified script
1
12,762
15,116,341,615
IssuesEvent
2021-02-09 06:33:41
yuta252/startlens_learning
https://api.github.com/repos/yuta252/startlens_learning
closed
S3 storageから学習用画像ファイルの取得
dev process
## 概要 DeepLearningモデルで学習するためには、StartlensアプリでS3に格納した画像ファイルを取得してモデルの入力層に渡す必要がある。そこでS3から画像を取得しハンドリングするためのclassを作成した。 ## 変更点 - フォルダ構成と初期設定 * settings.ini, settings.pyによる設定情報の管理 * Dockerfile, docker-compose.ymlによるインフラ環境の構築 * 環境変数.envファイルの設置 * constants.pyによる定数情報の管理 - fetch/resource.pyによるS3リソースのハンドリング ## 備考 - S3からの画像取得操作はboto3のboto3.resource(高レベルAPI)もしくはboto3.client(低レベルAPI)を利用する - S3から取得した学習用画像ファイルは、大容量になることが予想されるため、一時ファイルとして保存するのではなくByteIOを活用し、オンメモリ上で画像の加工編集を行う。 ## 課題 - オンメモリ上での画像処理と学習モデルへのインプットを行うが、学習を進めていくとメモリが逼迫する可能性がある。 - そこで、明示的にガベージコレクションを利用するなどしてメモリ管理の観点からの改良を図る。 ## 参照 - [boto3公式ドキュメント](https://boto3.amazonaws.com/v1/documentation/api/latest/guide/s3-example-download-file.html) - [boto3の認証情報検索順序](https://qiita.com/tsukamoto/items/00ec8ef7e9a4ce4fb0e9) - [boto3 で S3 から指定した prefix のオブジェクトをダウンロードする](https://rriifftt.hatenablog.com/entry/2018/06/04/144906) - [Python の boto3 で S3 とダウンロード/アップロードする](https://sig9.hatenablog.com/entry/2020/02/02/000000)
1.0
S3 storageから学習用画像ファイルの取得 - ## 概要 DeepLearningモデルで学習するためには、StartlensアプリでS3に格納した画像ファイルを取得してモデルの入力層に渡す必要がある。そこでS3から画像を取得しハンドリングするためのclassを作成した。 ## 変更点 - フォルダ構成と初期設定 * settings.ini, settings.pyによる設定情報の管理 * Dockerfile, docker-compose.ymlによるインフラ環境の構築 * 環境変数.envファイルの設置 * constants.pyによる定数情報の管理 - fetch/resource.pyによるS3リソースのハンドリング ## 備考 - S3からの画像取得操作はboto3のboto3.resource(高レベルAPI)もしくはboto3.client(低レベルAPI)を利用する - S3から取得した学習用画像ファイルは、大容量になることが予想されるため、一時ファイルとして保存するのではなくByteIOを活用し、オンメモリ上で画像の加工編集を行う。 ## 課題 - オンメモリ上での画像処理と学習モデルへのインプットを行うが、学習を進めていくとメモリが逼迫する可能性がある。 - そこで、明示的にガベージコレクションを利用するなどしてメモリ管理の観点からの改良を図る。 ## 参照 - [boto3公式ドキュメント](https://boto3.amazonaws.com/v1/documentation/api/latest/guide/s3-example-download-file.html) - [boto3の認証情報検索順序](https://qiita.com/tsukamoto/items/00ec8ef7e9a4ce4fb0e9) - [boto3 で S3 から指定した prefix のオブジェクトをダウンロードする](https://rriifftt.hatenablog.com/entry/2018/06/04/144906) - [Python の boto3 で S3 とダウンロード/アップロードする](https://sig9.hatenablog.com/entry/2020/02/02/000000)
process
storageから学習用画像ファイルの取得 概要 deeplearningモデルで学習するためには、 。 。 変更点 フォルダ構成と初期設定 settings ini settings pyによる設定情報の管理 dockerfile docker compose ymlによるインフラ環境の構築 環境変数 envファイルの設置 constants pyによる定数情報の管理 fetch resource 備考 resource 高レベルapi) client(低レベルapi)を利用する 、大容量になることが予想されるため、一時ファイルとして保存するのではなくbyteioを活用し、オンメモリ上で画像の加工編集を行う。 課題 オンメモリ上での画像処理と学習モデルへのインプットを行うが、学習を進めていくとメモリが逼迫する可能性がある。 そこで、明示的にガベージコレクションを利用するなどしてメモリ管理の観点からの改良を図る。 参照
1
2,991
5,968,535,226
IssuesEvent
2017-05-30 18:19:08
IFPB-2017-1/seminario
https://api.github.com/repos/IFPB-2017-1/seminario
closed
Criar TAP - Termo de abertura de projeto
Concluido Em processo
Seguir o modelo que está no Google Drive da professora para criar o documento inicial do projeto. Ele conterá as atribuições de cada membro do grupo e será o documento de visão com a descrição do projeto.
1.0
Criar TAP - Termo de abertura de projeto - Seguir o modelo que está no Google Drive da professora para criar o documento inicial do projeto. Ele conterá as atribuições de cada membro do grupo e será o documento de visão com a descrição do projeto.
process
criar tap termo de abertura de projeto seguir o modelo que está no google drive da professora para criar o documento inicial do projeto ele conterá as atribuições de cada membro do grupo e será o documento de visão com a descrição do projeto
1
19,719
10,419,850,159
IssuesEvent
2019-09-15 19:36:15
andOTP/andOTP
https://api.github.com/repos/andOTP/andOTP
closed
Bad crypto implementation?
question security
I just saw the comments in this [Reddit thread](https://old.reddit.com/r/androidapps/comments/b45zrj/dev_aegis_authenticator_secure_two_factor/ejvioko/?context=2). 1. Have the bad cryptography designs been addressed in the latest version? 2. If they havnt, have they been at least documented somewhere?
True
Bad crypto implementation? - I just saw the comments in this [Reddit thread](https://old.reddit.com/r/androidapps/comments/b45zrj/dev_aegis_authenticator_secure_two_factor/ejvioko/?context=2). 1. Have the bad cryptography designs been addressed in the latest version? 2. If they havnt, have they been at least documented somewhere?
non_process
bad crypto implementation i just saw the comments in this have the bad cryptography designs been addressed in the latest version if they havnt have they been at least documented somewhere
0
20,521
27,180,326,531
IssuesEvent
2023-02-18 14:48:36
OpenDataScotland/the_od_bods
https://api.github.com/repos/OpenDataScotland/the_od_bods
opened
Create alternative .csv output of dataset listing for end users
good first issue data processing front end
**Is your feature request related to a problem? Please describe.** Currently, we offer our dataset listing in a .csv and .json download. The .csv was the original output of merge_data.py but is now being replaced by a .json file itself. This means we lose the .csv format for public users. **Describe the solution you'd like** Create and a .csv format of the dataset listing that users can download. **Describe alternatives you've considered** We've not considered where the .csv should live, so this is up for discussion. **Additional context** Original ticket triggering this change is #163
1.0
Create alternative .csv output of dataset listing for end users - **Is your feature request related to a problem? Please describe.** Currently, we offer our dataset listing in a .csv and .json download. The .csv was the original output of merge_data.py but is now being replaced by a .json file itself. This means we lose the .csv format for public users. **Describe the solution you'd like** Create and a .csv format of the dataset listing that users can download. **Describe alternatives you've considered** We've not considered where the .csv should live, so this is up for discussion. **Additional context** Original ticket triggering this change is #163
process
create alternative csv output of dataset listing for end users is your feature request related to a problem please describe currently we offer our dataset listing in a csv and json download the csv was the original output of merge data py but is now being replaced by a json file itself this means we lose the csv format for public users describe the solution you d like create and a csv format of the dataset listing that users can download describe alternatives you ve considered we ve not considered where the csv should live so this is up for discussion additional context original ticket triggering this change is
1
231,045
17,661,016,988
IssuesEvent
2021-08-21 14:02:12
borgbackup/borg
https://api.github.com/repos/borgbackup/borg
closed
borg list patterns doc / example
documentation
I was trying to use `borg list` with path patterns but the documentation was lacking. I didn't see how to do it until I found a email thread with an example. The docs only say: >   | PATH | paths to list; patterns are supported My understanding is that unlike with `--exclude`, patterns are not recognized by default, but require a prefix like `re:`. It would be nice to have at least one example and maybe a line that says something like: > if `path` starts with a pattern prefix, it will be treated like a pattern. Otherwise it must be an exact match."
1.0
borg list patterns doc / example - I was trying to use `borg list` with path patterns but the documentation was lacking. I didn't see how to do it until I found a email thread with an example. The docs only say: >   | PATH | paths to list; patterns are supported My understanding is that unlike with `--exclude`, patterns are not recognized by default, but require a prefix like `re:`. It would be nice to have at least one example and maybe a line that says something like: > if `path` starts with a pattern prefix, it will be treated like a pattern. Otherwise it must be an exact match."
non_process
borg list patterns doc example i was trying to use borg list with path patterns but the documentation was lacking i didn t see how to do it until i found a email thread with an example the docs only say   path paths to list patterns are supported my understanding is that unlike with exclude patterns are not recognized by default but require a prefix like re it would be nice to have at least one example and maybe a line that says something like if path starts with a pattern prefix it will be treated like a pattern otherwise it must be an exact match
0
170,189
13,177,646,772
IssuesEvent
2020-08-12 07:46:00
microsoft/AzureStorageExplorer
https://api.github.com/repos/microsoft/AzureStorageExplorer
closed
No error information displays when typing other service's URL in 'Connect to public blob container' dialog
:gear: blobs 🧪 testing
**Storage Explorer Version:** 1.15.0-dev **Build**: 20200717.1 **Branch**: master **Platform/OS:** Windows 10/ Linux Ubuntu 18.04/ macOS Catalina **Architecture**: ia32/x64 **Regression From:** Not a regression **Steps to reproduce:** 1. Expand one storage account -> Copy the URL of one queue. 2. Open connect dialog -> Select 'Connect to public blob container' -> Paste the URL to 'Container URL' box in the dialog. 3. Check the result. **Expect Experience:** An error information displays. ![image](https://user-images.githubusercontent.com/54055206/87768463-be10d580-c84e-11ea-8cd3-f9374d8b9348.png) **Actual Experience:** No error information displays. ![image](https://user-images.githubusercontent.com/54055206/87768497-cbc65b00-c84e-11ea-8c55-a2b950c3593a.png) **More Info:** This issue also reproduces for file shares/tables.
1.0
No error information displays when typing other service's URL in 'Connect to public blob container' dialog - **Storage Explorer Version:** 1.15.0-dev **Build**: 20200717.1 **Branch**: master **Platform/OS:** Windows 10/ Linux Ubuntu 18.04/ macOS Catalina **Architecture**: ia32/x64 **Regression From:** Not a regression **Steps to reproduce:** 1. Expand one storage account -> Copy the URL of one queue. 2. Open connect dialog -> Select 'Connect to public blob container' -> Paste the URL to 'Container URL' box in the dialog. 3. Check the result. **Expect Experience:** An error information displays. ![image](https://user-images.githubusercontent.com/54055206/87768463-be10d580-c84e-11ea-8cd3-f9374d8b9348.png) **Actual Experience:** No error information displays. ![image](https://user-images.githubusercontent.com/54055206/87768497-cbc65b00-c84e-11ea-8c55-a2b950c3593a.png) **More Info:** This issue also reproduces for file shares/tables.
non_process
no error information displays when typing other service s url in connect to public blob container dialog storage explorer version dev build branch master platform os windows linux ubuntu macos catalina architecture regression from not a regression steps to reproduce expand one storage account copy the url of one queue open connect dialog select connect to public blob container paste the url to container url box in the dialog check the result expect experience an error information displays actual experience no error information displays more info this issue also reproduces for file shares tables
0
7,176
10,318,672,013
IssuesEvent
2019-08-30 15:28:54
prisma/prisma2
https://api.github.com/repos/prisma/prisma2
closed
Examples and Docs should not use both @id and @unique
kind/docs process/candidate
In a lot of bug reports i see `@id` and `@unique` being used in conjunction e.g.: ``` model User { id String @default(cuid()) @id @unique ... } ``` This is not necessary though. `@id` already implies that something is unique. Hence `@unique` is not needed. I think we should go through our docs and examples and remove it to avoid that this pattern proliferates further.
1.0
Examples and Docs should not use both @id and @unique - In a lot of bug reports i see `@id` and `@unique` being used in conjunction e.g.: ``` model User { id String @default(cuid()) @id @unique ... } ``` This is not necessary though. `@id` already implies that something is unique. Hence `@unique` is not needed. I think we should go through our docs and examples and remove it to avoid that this pattern proliferates further.
process
examples and docs should not use both id and unique in a lot of bug reports i see id and unique being used in conjunction e g model user id string default cuid id unique this is not necessary though id already implies that something is unique hence unique is not needed i think we should go through our docs and examples and remove it to avoid that this pattern proliferates further
1
398,193
11,739,063,328
IssuesEvent
2020-03-11 17:04:18
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
Cannot use mixed-type question filters on dashboard
Priority:P2 Querying/Parameters & Variables Reporting/Dashboards Type:Bug
**Describe the bug** When mixing and matching questions with different filter types (e.g. Field Filter vs. Number/Text Filter), I occasionally receive errors depending on how I load the dashboard. **Logs** ``` 02-10 18:27:49 WARN middleware.process-userland-query :: Query failure {:status :failed, :class clojure.lang.ExceptionInfo, :error "Output of value->number does not match schema: \n\n\t (not (matches-some-precondition? nil)) \n\n", :stacktrace ("--> driver.common.parameters.values$fn__72804$query__GT_params_map__72809$fn__72813.invoke(values.clj:235)" "driver.common.parameters.values$fn__72804$query__GT_params_map__72809.invoke(values.clj:220)" "driver.sql$fn__74176.invokeStatic(sql.clj:42)" "driver.sql$fn__74176.invoke(sql.clj:38)" "query_processor.middleware.parameters.native$expand_inner.invokeStatic(native.clj:39)" "query_processor.middleware.parameters.native$expand_inner.invoke(native.clj:30)" "query_processor.middleware.parameters$expand_one.invokeStatic(parameters.clj:50)" "query_processor.middleware.parameters$expand_one.invoke(parameters.clj:41)" "query_processor.middleware.parameters$expand_all$replace_44425__44426.invoke(parameters.clj:59)" "mbql.util.match$replace_in_collection$iter__26259__26263$fn__26264.invoke(match.clj:132)" "mbql.util.match$replace_in_collection.invokeStatic(match.clj:131)" "mbql.util.match$replace_in_collection.invoke(match.clj:126)" "query_processor.middleware.parameters$expand_all$replace_44425__44426.invoke(parameters.clj:59)" "query_processor.middleware.parameters$expand_all.invokeStatic(parameters.clj:59)" "query_processor.middleware.parameters$expand_all.invoke(parameters.clj:53)" "query_processor.middleware.parameters$expand_all.invokeStatic(parameters.clj:56)" "query_processor.middleware.parameters$expand_all.invoke(parameters.clj:53)" "query_processor.middleware.parameters$expand_parameters.invokeStatic(parameters.clj:77)" "query_processor.middleware.parameters$expand_parameters.invoke(parameters.clj:73)" "query_processor.middleware.parameters$fn__44441$substitute_parameters_STAR___44446$fn__44447.invoke(parameters.clj:82)" "query_processor.middleware.parameters$fn__44441$substitute_parameters_STAR___44446.invoke(parameters.clj:79)" "query_processor.middleware.driver_specific$process_query_in_context$fn__43375.invoke(driver_specific.clj:12)" "query_processor.middleware.resolve_driver$resolve_driver$fn__44774.invoke(resolve_driver.clj:22)" "query_processor.middleware.store$initialize_store$fn__47900$fn__47901.invoke(store.clj:11)" "query_processor.store$do_with_store.invokeStatic(store.clj:46)" "query_processor.store$do_with_store.invoke(store.clj:40)" "query_processor.middleware.store$initialize_store$fn__47900.invoke(store.clj:10)" "query_processor.middleware.async$async__GT_sync$fn__40271.invoke(async.clj:23)" "query_processor.middleware.async_wait$runnable$fn__42416.invoke(async_wait.clj:89)"), :query {:constraints {:max-results 10000, :max-results-bare-rows 2000}, :type :native, :middleware {:userland-query? true}, :native {:query "SELECT * FROM ORDERS WHERE ID = {{seller_id}}", :template-tags {"seller_id" {:id "c26ea0ea-4972-fabc-859b-919b2120926a", :name "seller_id", :display-name "Seller ID", :type :number, :default "2", :required true}}}, :info {:executed-by 1, :context :question, :card-id 28, :dashboard-id nil, :query-hash [-89, 90, -6, 117, -80, 81, -118, 122, 11, -32, 121, -50, -122, 70, 40, 79, 70, 9, -5, 69, 40, 61, 122, 33, 10, 113, -83, -112, -66, -41, -70, 28]}, :parameters [{:type "category", :target ["variable" ["template-tag" "seller_id"]], :value ["2" "3"]}], :async? true, :cache-ttl nil}, :cause {:class clojure.lang.ExceptionInfo, :error (not (matches-some-precondition? nil)), :ex-data {:type :schema.core/error, :value nil, :error (not (matches-some-precondition? nil))}}, :ex-data {:type :schema.core/error, :tags {"seller_id" {:id "c26ea0ea-4972-fabc-859b-919b2120926a", :name "seller_id", :display-name "Seller ID", :type :number, :default "2", :required true}}, :params [{:type :category, :target [:variable [:template-tag "seller_id"]], :value ["2" "3"]}]}} ``` **To Reproduce** There are two different scenarios where I've seen this happen. 1. Mixed simple/native filter with text/number types 2. Mixed simple/native filter both with field filter types (Postgres) This issue is for the first, I'll file a second issue for the other. Steps to reproduce the behavior: 1. Create 3 new questions with sample dataset 2. 1: Simple question, summarize Orders 3. 2: Native question (`SELECT * FROM ORDERS WHERE {{seller_id}}`) with Field filter on Orders-->ID 4. 3. Native question (`SELECT * FROM ORDERS WHERE ID = {{seller_id}}`) with Number filter on Orders-->ID 5. Create a dashboard with those 3 three questions and a filter with a default value 6. Load the dashboard and notice that the native number question doesn't load re: step 6, sometimes depending on how you load the dashboard, it will work. If I browse to the dashboard in the app, it fails. But if I then reload that page, it loads successfully. If I add a second filter parameter (e.g. http://localhost:3000/dashboard/1?id=2&id=3 ), then it usually always fails. **Expected behavior** Questions on the dashboard load regardless of what type of filter I'm using. **Screenshots** n/a **Information about your Metabase Installation:** ```json { "browser-info": { "language": "en-US", "platform": "MacIntel", "userAgent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_14_6) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/80.0.3987.87 Safari/537.36", "vendor": "Google Inc." }, "system-info": { "java.runtime.name": "OpenJDK Runtime Environment", "java.runtime.version": "11.0.4+11", "java.vendor": "AdoptOpenJDK", "java.vendor.url": "https://adoptopenjdk.net/", "java.version": "11.0.4", "java.vm.name": "OpenJDK 64-Bit Server VM", "java.vm.version": "11.0.4+11", "os.name": "Linux", "os.version": "4.19.76-linuxkit", "user.language": "en", "user.timezone": "Etc/UTC" }, "metabase-info": { "databases": [ "postgres", "h2" ], "hosting-env": "unknown", "application-database": "postgres", "run-mode": "prod", "version": { "date": "2020-01-16", "tag": "v1.34.1", "branch": "enterprise-release-1.34.x", "hash": "a6df4f6" }, "settings": { "report-timezone": null } } } ``` **Severity** Annoying - I can fix it with a page reload in some cases. **Additional context** n/a
1.0
Cannot use mixed-type question filters on dashboard - **Describe the bug** When mixing and matching questions with different filter types (e.g. Field Filter vs. Number/Text Filter), I occasionally receive errors depending on how I load the dashboard. **Logs** ``` 02-10 18:27:49 WARN middleware.process-userland-query :: Query failure {:status :failed, :class clojure.lang.ExceptionInfo, :error "Output of value->number does not match schema: \n\n\t (not (matches-some-precondition? nil)) \n\n", :stacktrace ("--> driver.common.parameters.values$fn__72804$query__GT_params_map__72809$fn__72813.invoke(values.clj:235)" "driver.common.parameters.values$fn__72804$query__GT_params_map__72809.invoke(values.clj:220)" "driver.sql$fn__74176.invokeStatic(sql.clj:42)" "driver.sql$fn__74176.invoke(sql.clj:38)" "query_processor.middleware.parameters.native$expand_inner.invokeStatic(native.clj:39)" "query_processor.middleware.parameters.native$expand_inner.invoke(native.clj:30)" "query_processor.middleware.parameters$expand_one.invokeStatic(parameters.clj:50)" "query_processor.middleware.parameters$expand_one.invoke(parameters.clj:41)" "query_processor.middleware.parameters$expand_all$replace_44425__44426.invoke(parameters.clj:59)" "mbql.util.match$replace_in_collection$iter__26259__26263$fn__26264.invoke(match.clj:132)" "mbql.util.match$replace_in_collection.invokeStatic(match.clj:131)" "mbql.util.match$replace_in_collection.invoke(match.clj:126)" "query_processor.middleware.parameters$expand_all$replace_44425__44426.invoke(parameters.clj:59)" "query_processor.middleware.parameters$expand_all.invokeStatic(parameters.clj:59)" "query_processor.middleware.parameters$expand_all.invoke(parameters.clj:53)" "query_processor.middleware.parameters$expand_all.invokeStatic(parameters.clj:56)" "query_processor.middleware.parameters$expand_all.invoke(parameters.clj:53)" "query_processor.middleware.parameters$expand_parameters.invokeStatic(parameters.clj:77)" "query_processor.middleware.parameters$expand_parameters.invoke(parameters.clj:73)" "query_processor.middleware.parameters$fn__44441$substitute_parameters_STAR___44446$fn__44447.invoke(parameters.clj:82)" "query_processor.middleware.parameters$fn__44441$substitute_parameters_STAR___44446.invoke(parameters.clj:79)" "query_processor.middleware.driver_specific$process_query_in_context$fn__43375.invoke(driver_specific.clj:12)" "query_processor.middleware.resolve_driver$resolve_driver$fn__44774.invoke(resolve_driver.clj:22)" "query_processor.middleware.store$initialize_store$fn__47900$fn__47901.invoke(store.clj:11)" "query_processor.store$do_with_store.invokeStatic(store.clj:46)" "query_processor.store$do_with_store.invoke(store.clj:40)" "query_processor.middleware.store$initialize_store$fn__47900.invoke(store.clj:10)" "query_processor.middleware.async$async__GT_sync$fn__40271.invoke(async.clj:23)" "query_processor.middleware.async_wait$runnable$fn__42416.invoke(async_wait.clj:89)"), :query {:constraints {:max-results 10000, :max-results-bare-rows 2000}, :type :native, :middleware {:userland-query? true}, :native {:query "SELECT * FROM ORDERS WHERE ID = {{seller_id}}", :template-tags {"seller_id" {:id "c26ea0ea-4972-fabc-859b-919b2120926a", :name "seller_id", :display-name "Seller ID", :type :number, :default "2", :required true}}}, :info {:executed-by 1, :context :question, :card-id 28, :dashboard-id nil, :query-hash [-89, 90, -6, 117, -80, 81, -118, 122, 11, -32, 121, -50, -122, 70, 40, 79, 70, 9, -5, 69, 40, 61, 122, 33, 10, 113, -83, -112, -66, -41, -70, 28]}, :parameters [{:type "category", :target ["variable" ["template-tag" "seller_id"]], :value ["2" "3"]}], :async? true, :cache-ttl nil}, :cause {:class clojure.lang.ExceptionInfo, :error (not (matches-some-precondition? nil)), :ex-data {:type :schema.core/error, :value nil, :error (not (matches-some-precondition? nil))}}, :ex-data {:type :schema.core/error, :tags {"seller_id" {:id "c26ea0ea-4972-fabc-859b-919b2120926a", :name "seller_id", :display-name "Seller ID", :type :number, :default "2", :required true}}, :params [{:type :category, :target [:variable [:template-tag "seller_id"]], :value ["2" "3"]}]}} ``` **To Reproduce** There are two different scenarios where I've seen this happen. 1. Mixed simple/native filter with text/number types 2. Mixed simple/native filter both with field filter types (Postgres) This issue is for the first, I'll file a second issue for the other. Steps to reproduce the behavior: 1. Create 3 new questions with sample dataset 2. 1: Simple question, summarize Orders 3. 2: Native question (`SELECT * FROM ORDERS WHERE {{seller_id}}`) with Field filter on Orders-->ID 4. 3. Native question (`SELECT * FROM ORDERS WHERE ID = {{seller_id}}`) with Number filter on Orders-->ID 5. Create a dashboard with those 3 three questions and a filter with a default value 6. Load the dashboard and notice that the native number question doesn't load re: step 6, sometimes depending on how you load the dashboard, it will work. If I browse to the dashboard in the app, it fails. But if I then reload that page, it loads successfully. If I add a second filter parameter (e.g. http://localhost:3000/dashboard/1?id=2&id=3 ), then it usually always fails. **Expected behavior** Questions on the dashboard load regardless of what type of filter I'm using. **Screenshots** n/a **Information about your Metabase Installation:** ```json { "browser-info": { "language": "en-US", "platform": "MacIntel", "userAgent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_14_6) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/80.0.3987.87 Safari/537.36", "vendor": "Google Inc." }, "system-info": { "java.runtime.name": "OpenJDK Runtime Environment", "java.runtime.version": "11.0.4+11", "java.vendor": "AdoptOpenJDK", "java.vendor.url": "https://adoptopenjdk.net/", "java.version": "11.0.4", "java.vm.name": "OpenJDK 64-Bit Server VM", "java.vm.version": "11.0.4+11", "os.name": "Linux", "os.version": "4.19.76-linuxkit", "user.language": "en", "user.timezone": "Etc/UTC" }, "metabase-info": { "databases": [ "postgres", "h2" ], "hosting-env": "unknown", "application-database": "postgres", "run-mode": "prod", "version": { "date": "2020-01-16", "tag": "v1.34.1", "branch": "enterprise-release-1.34.x", "hash": "a6df4f6" }, "settings": { "report-timezone": null } } } ``` **Severity** Annoying - I can fix it with a page reload in some cases. **Additional context** n/a
non_process
cannot use mixed type question filters on dashboard describe the bug when mixing and matching questions with different filter types e g field filter vs number text filter i occasionally receive errors depending on how i load the dashboard logs warn middleware process userland query query failure status failed class clojure lang exceptioninfo error output of value number does not match schema n n t not matches some precondition nil n n stacktrace driver common parameters values fn query gt params map fn invoke values clj driver common parameters values fn query gt params map invoke values clj driver sql fn invokestatic sql clj driver sql fn invoke sql clj query processor middleware parameters native expand inner invokestatic native clj query processor middleware parameters native expand inner invoke native clj query processor middleware parameters expand one invokestatic parameters clj query processor middleware parameters expand one invoke parameters clj query processor middleware parameters expand all replace invoke parameters clj mbql util match replace in collection iter fn invoke match clj mbql util match replace in collection invokestatic match clj mbql util match replace in collection invoke match clj query processor middleware parameters expand all replace invoke parameters clj query processor middleware parameters expand all invokestatic parameters clj query processor middleware parameters expand all invoke parameters clj query processor middleware parameters expand all invokestatic parameters clj query processor middleware parameters expand all invoke parameters clj query processor middleware parameters expand parameters invokestatic parameters clj query processor middleware parameters expand parameters invoke parameters clj query processor middleware parameters fn substitute parameters star fn invoke parameters clj query processor middleware parameters fn substitute parameters star invoke parameters clj query processor middleware driver specific process query in context fn invoke driver specific clj query processor middleware resolve driver resolve driver fn invoke resolve driver clj query processor middleware store initialize store fn fn invoke store clj query processor store do with store invokestatic store clj query processor store do with store invoke store clj query processor middleware store initialize store fn invoke store clj query processor middleware async async gt sync fn invoke async clj query processor middleware async wait runnable fn invoke async wait clj query constraints max results max results bare rows type native middleware userland query true native query select from orders where id seller id template tags seller id id fabc name seller id display name seller id type number default required true info executed by context question card id dashboard id nil query hash parameters value async true cache ttl nil cause class clojure lang exceptioninfo error not matches some precondition nil ex data type schema core error value nil error not matches some precondition nil ex data type schema core error tags seller id id fabc name seller id display name seller id type number default required true params value to reproduce there are two different scenarios where i ve seen this happen mixed simple native filter with text number types mixed simple native filter both with field filter types postgres this issue is for the first i ll file a second issue for the other steps to reproduce the behavior create new questions with sample dataset simple question summarize orders native question select from orders where seller id with field filter on orders id native question select from orders where id seller id with number filter on orders id create a dashboard with those three questions and a filter with a default value load the dashboard and notice that the native number question doesn t load re step sometimes depending on how you load the dashboard it will work if i browse to the dashboard in the app it fails but if i then reload that page it loads successfully if i add a second filter parameter e g then it usually always fails expected behavior questions on the dashboard load regardless of what type of filter i m using screenshots n a information about your metabase installation json browser info language en us platform macintel useragent mozilla macintosh intel mac os x applewebkit khtml like gecko chrome safari vendor google inc system info java runtime name openjdk runtime environment java runtime version java vendor adoptopenjdk java vendor url java version java vm name openjdk bit server vm java vm version os name linux os version linuxkit user language en user timezone etc utc metabase info databases postgres hosting env unknown application database postgres run mode prod version date tag branch enterprise release x hash settings report timezone null severity annoying i can fix it with a page reload in some cases additional context n a
0
280,339
24,296,338,240
IssuesEvent
2022-09-29 10:21:31
wpfoodmanager/wp-food-manager
https://api.github.com/repos/wpfoodmanager/wp-food-manager
closed
Backend - Able to change food postion
In Testing
Able to change food postion . This should not be change. https://user-images.githubusercontent.com/75515088/192772946-8ff3c4b0-f19d-4f38-83f1-d7b8c5e0c6f5.mp4
1.0
Backend - Able to change food postion - Able to change food postion . This should not be change. https://user-images.githubusercontent.com/75515088/192772946-8ff3c4b0-f19d-4f38-83f1-d7b8c5e0c6f5.mp4
non_process
backend able to change food postion able to change food postion this should not be change
0
23,740
16,550,549,680
IssuesEvent
2021-05-28 08:05:14
google/site-kit-wp
https://api.github.com/repos/google/site-kit-wp
opened
Add a security policy
P1 Type: Infrastructure
## Feature Description https://docs.github.com/en/code-security/getting-started/adding-a-security-policy-to-your-repository Eg. https://github.com/ampproject/amp-wp/blob/3bb15f3d660c4401b958147430664d07eb1e03e2/SECURITY.md --------------- _Do not alter or remove anything below. The following sections will be managed by moderators only._ ## Acceptance criteria * <!-- One or more bullet points for acceptance criteria. --> ## Implementation Brief * <!-- One or more bullet points for how to technically implement the feature. --> ### Test Coverage * <!-- One or more bullet points for how to implement automated tests to verify the feature works. --> ### Visual Regression Changes * <!-- One or more bullet points describing how the feature will affect visual regression tests, if applicable. --> ## QA Brief * <!-- One or more bullet points for how to test that the feature works as expected. --> ## Changelog entry * <!-- One sentence summarizing the PR, to be used in the changelog. -->
1.0
Add a security policy - ## Feature Description https://docs.github.com/en/code-security/getting-started/adding-a-security-policy-to-your-repository Eg. https://github.com/ampproject/amp-wp/blob/3bb15f3d660c4401b958147430664d07eb1e03e2/SECURITY.md --------------- _Do not alter or remove anything below. The following sections will be managed by moderators only._ ## Acceptance criteria * <!-- One or more bullet points for acceptance criteria. --> ## Implementation Brief * <!-- One or more bullet points for how to technically implement the feature. --> ### Test Coverage * <!-- One or more bullet points for how to implement automated tests to verify the feature works. --> ### Visual Regression Changes * <!-- One or more bullet points describing how the feature will affect visual regression tests, if applicable. --> ## QA Brief * <!-- One or more bullet points for how to test that the feature works as expected. --> ## Changelog entry * <!-- One sentence summarizing the PR, to be used in the changelog. -->
non_process
add a security policy feature description eg do not alter or remove anything below the following sections will be managed by moderators only acceptance criteria implementation brief test coverage visual regression changes qa brief changelog entry
0
40,949
10,238,386,188
IssuesEvent
2019-08-19 15:46:16
PowerDNS/pdns
https://api.github.com/repos/PowerDNS/pdns
opened
auth nsupdate: duplicate entries when mixing case
auth defect
- Program: Authoritative - Issue type: Bug report/Feature request ### Short description Our 2136-code prevents adding duplicate entries, but this fails in the face of mixed case entries. Patching one of the existing tests like this reveals the issue: ```diff diff --git a/regression-tests/tests/1dyndns-update-replace-cname/command b/regression-tests/tests/1dyndns-update-replace-cname/command index 17bd7c818..3eef41f0e 100755 --- a/regression-tests/tests/1dyndns-update-replace-cname/command +++ b/regression-tests/tests/1dyndns-update-replace-cname/command @@ -1,22 +1,22 @@ #!/bin/sh -cleandig cname1.test.dyndns CNAME +cleandig ptr1.test.dyndns PTR cleannsupdate <<! server $nameserver $port zone test.dyndns -update add cname1.test.dyndns. 3600 CNAME host-2.test.dyndns. +update add ptr1.test.dyndns. 3600 PTR host-2.test.dyndns. send answer ! -cleandig cname1.test.dyndns CNAME +cleandig ptr1.test.dyndns PTR cleannsupdate <<! server $nameserver $port zone test.dyndns -update add cname1.test.dyndns. 3600 CNAME host-1.test.dyndns. +update add ptr1.test.dyndns. 3600 PTR HOST-2.test.dyndns. send answer ! -cleandig cname1.test.dyndns CNAME +cleandig ptr1.test.dyndns PTR ```
1.0
auth nsupdate: duplicate entries when mixing case - - Program: Authoritative - Issue type: Bug report/Feature request ### Short description Our 2136-code prevents adding duplicate entries, but this fails in the face of mixed case entries. Patching one of the existing tests like this reveals the issue: ```diff diff --git a/regression-tests/tests/1dyndns-update-replace-cname/command b/regression-tests/tests/1dyndns-update-replace-cname/command index 17bd7c818..3eef41f0e 100755 --- a/regression-tests/tests/1dyndns-update-replace-cname/command +++ b/regression-tests/tests/1dyndns-update-replace-cname/command @@ -1,22 +1,22 @@ #!/bin/sh -cleandig cname1.test.dyndns CNAME +cleandig ptr1.test.dyndns PTR cleannsupdate <<! server $nameserver $port zone test.dyndns -update add cname1.test.dyndns. 3600 CNAME host-2.test.dyndns. +update add ptr1.test.dyndns. 3600 PTR host-2.test.dyndns. send answer ! -cleandig cname1.test.dyndns CNAME +cleandig ptr1.test.dyndns PTR cleannsupdate <<! server $nameserver $port zone test.dyndns -update add cname1.test.dyndns. 3600 CNAME host-1.test.dyndns. +update add ptr1.test.dyndns. 3600 PTR HOST-2.test.dyndns. send answer ! -cleandig cname1.test.dyndns CNAME +cleandig ptr1.test.dyndns PTR ```
non_process
auth nsupdate duplicate entries when mixing case program authoritative issue type bug report feature request short description our code prevents adding duplicate entries but this fails in the face of mixed case entries patching one of the existing tests like this reveals the issue diff diff git a regression tests tests update replace cname command b regression tests tests update replace cname command index a regression tests tests update replace cname command b regression tests tests update replace cname command bin sh cleandig test dyndns cname cleandig test dyndns ptr cleannsupdate server nameserver port zone test dyndns update add test dyndns cname host test dyndns update add test dyndns ptr host test dyndns send answer cleandig test dyndns cname cleandig test dyndns ptr cleannsupdate server nameserver port zone test dyndns update add test dyndns cname host test dyndns update add test dyndns ptr host test dyndns send answer cleandig test dyndns cname cleandig test dyndns ptr
0
6,182
9,100,315,465
IssuesEvent
2019-02-20 08:11:00
comic/grand-challenge.org
https://api.github.com/repos/comic/grand-challenge.org
opened
Replace mhd/zraw with mha as internal image representation?
area/ophthalmology-workstation area/processors
I think that we should replace mhd/zraw with compressed mha as our internal representation. I think that this would have the following advantages: - At the moment, we re-write all meta image files to be named out.mhd and out.zraw, which are located at `images/<image.pk>/out.<type>`. This is awkward as everywhere we need to handle two files, sort out which one is the header, and already in the codebase there are a lot of select by file extension statements. - If we want to share a bunch of files, say, as an ImageSet for training, all of the files will have the same name, and will have to be re-written (including mhd modification) before they're given to the user. I would prefer that we use mha files. Then, we only have to deal with 1 file, we don't need to sort it everywhere, and we can rename it on the fly. This would then be consistent with the tiff representation. We could then serve images from `images/<image.pk>/<imagefile.pk>/`. I'm fairly sure that this would still be compatible with the Python (SimpleITK), Mevislab and Javascript (ITK.js) loading libraries. What could break: - Everywhere that we've made an assumption about getting 2 images: - CIRRUS Web - The `get_sitk_image` function (mhd is not used directly by the ophthalmology workstation). - Processors: how do the existing algorithms handle mhd? Any thoughts? cc. @pkcakeout @HarmvZ
1.0
Replace mhd/zraw with mha as internal image representation? - I think that we should replace mhd/zraw with compressed mha as our internal representation. I think that this would have the following advantages: - At the moment, we re-write all meta image files to be named out.mhd and out.zraw, which are located at `images/<image.pk>/out.<type>`. This is awkward as everywhere we need to handle two files, sort out which one is the header, and already in the codebase there are a lot of select by file extension statements. - If we want to share a bunch of files, say, as an ImageSet for training, all of the files will have the same name, and will have to be re-written (including mhd modification) before they're given to the user. I would prefer that we use mha files. Then, we only have to deal with 1 file, we don't need to sort it everywhere, and we can rename it on the fly. This would then be consistent with the tiff representation. We could then serve images from `images/<image.pk>/<imagefile.pk>/`. I'm fairly sure that this would still be compatible with the Python (SimpleITK), Mevislab and Javascript (ITK.js) loading libraries. What could break: - Everywhere that we've made an assumption about getting 2 images: - CIRRUS Web - The `get_sitk_image` function (mhd is not used directly by the ophthalmology workstation). - Processors: how do the existing algorithms handle mhd? Any thoughts? cc. @pkcakeout @HarmvZ
process
replace mhd zraw with mha as internal image representation i think that we should replace mhd zraw with compressed mha as our internal representation i think that this would have the following advantages at the moment we re write all meta image files to be named out mhd and out zraw which are located at images out this is awkward as everywhere we need to handle two files sort out which one is the header and already in the codebase there are a lot of select by file extension statements if we want to share a bunch of files say as an imageset for training all of the files will have the same name and will have to be re written including mhd modification before they re given to the user i would prefer that we use mha files then we only have to deal with file we don t need to sort it everywhere and we can rename it on the fly this would then be consistent with the tiff representation we could then serve images from images i m fairly sure that this would still be compatible with the python simpleitk mevislab and javascript itk js loading libraries what could break everywhere that we ve made an assumption about getting images cirrus web the get sitk image function mhd is not used directly by the ophthalmology workstation processors how do the existing algorithms handle mhd any thoughts cc pkcakeout harmvz
1
14,546
17,662,984,241
IssuesEvent
2021-08-21 22:32:52
GSG-FC03/adnan-Tic-Tac-Toe
https://api.github.com/repos/GSG-FC03/adnan-Tic-Tac-Toe
opened
functionality
in-process T 5hr
Creating functions that make the Tic Tac Toe works - [ ] Place the mark - [ ] check for winner - [ ] check for draw - [ ] switch turns
1.0
functionality - Creating functions that make the Tic Tac Toe works - [ ] Place the mark - [ ] check for winner - [ ] check for draw - [ ] switch turns
process
functionality creating functions that make the tic tac toe works place the mark check for winner check for draw switch turns
1
1,151
3,066,874,863
IssuesEvent
2015-08-18 06:32:23
TeamMentor/TM_4_0_Design
https://api.github.com/repos/TeamMentor/TM_4_0_Design
closed
Map out current security practices with 'Simplified Implementation of the Microsoft SDL'
Area: Security P3 Type: Task
* http://www.microsoft.com/security/sdl/default.aspx * docs can be downloaded from http://www.microsoft.com/en-us/download/details.aspx?id=12379 ![image](https://cloud.githubusercontent.com/assets/656739/5792987/ef319ac2-9f29-11e4-94a0-39dae79a11cc.png) We are already doing a lot of these, but it will be good to formalize them (for example on the one SDL steps that is missing is [Create a Threat Model for TM 4.0 (Jade, Flare and GraphDB)](https://github.com/TeamMentor/TM_4_0_Design/issues/269) ) ---- [TM-4.0-Security](https://github.com/TeamMentor/TM_4_0_Design/wiki/TM-4.0-Security)
True
Map out current security practices with 'Simplified Implementation of the Microsoft SDL' - * http://www.microsoft.com/security/sdl/default.aspx * docs can be downloaded from http://www.microsoft.com/en-us/download/details.aspx?id=12379 ![image](https://cloud.githubusercontent.com/assets/656739/5792987/ef319ac2-9f29-11e4-94a0-39dae79a11cc.png) We are already doing a lot of these, but it will be good to formalize them (for example on the one SDL steps that is missing is [Create a Threat Model for TM 4.0 (Jade, Flare and GraphDB)](https://github.com/TeamMentor/TM_4_0_Design/issues/269) ) ---- [TM-4.0-Security](https://github.com/TeamMentor/TM_4_0_Design/wiki/TM-4.0-Security)
non_process
map out current security practices with simplified implementation of the microsoft sdl docs can be downloaded from we are already doing a lot of these but it will be good to formalize them for example on the one sdl steps that is missing is
0
37,475
8,301,518,280
IssuesEvent
2018-09-21 11:44:44
Altinn/altinn-studio
https://api.github.com/repos/Altinn/altinn-studio
opened
As a service developer I should be able to add translations for codelist values
codelist
Kunne legge inn tekstnøkler Apiet skal bytte ut tekstnøkler med tekst
1.0
As a service developer I should be able to add translations for codelist values - Kunne legge inn tekstnøkler Apiet skal bytte ut tekstnøkler med tekst
non_process
as a service developer i should be able to add translations for codelist values kunne legge inn tekstnøkler apiet skal bytte ut tekstnøkler med tekst
0
55,441
30,753,416,195
IssuesEvent
2023-07-28 21:53:47
reportportal/reportportal
https://api.github.com/repos/reportportal/reportportal
closed
Unknown 'Auto-analysis' end status.
bug Check: Performance
**Describe the bug** 1. Open Launches page 2. Open 'More' menu for particular launch 3. Choose 'Analysis' 4. Auto-analysis started **Actual behavior** 'Auto-analysis' icon appears and constantly flashing all time. There is no any provided status of analysis. How should I recognise when it finishes? **Expected behavior** The end status of auto-analysis is displayed. **Screenshots** ![Screenshot 2023-06-26 at 16 12 11](https://github.com/reportportal/reportportal/assets/11822634/3acb75e0-7c84-4a21-abb6-83b1c0f30475) **Versions:** - macOS 13.3.1 (a), Chrome 114.0.5735.133 - Version of RP 23.1
True
Unknown 'Auto-analysis' end status. - **Describe the bug** 1. Open Launches page 2. Open 'More' menu for particular launch 3. Choose 'Analysis' 4. Auto-analysis started **Actual behavior** 'Auto-analysis' icon appears and constantly flashing all time. There is no any provided status of analysis. How should I recognise when it finishes? **Expected behavior** The end status of auto-analysis is displayed. **Screenshots** ![Screenshot 2023-06-26 at 16 12 11](https://github.com/reportportal/reportportal/assets/11822634/3acb75e0-7c84-4a21-abb6-83b1c0f30475) **Versions:** - macOS 13.3.1 (a), Chrome 114.0.5735.133 - Version of RP 23.1
non_process
unknown auto analysis end status describe the bug open launches page open more menu for particular launch choose analysis auto analysis started actual behavior auto analysis icon appears and constantly flashing all time there is no any provided status of analysis how should i recognise when it finishes expected behavior the end status of auto analysis is displayed screenshots versions macos a chrome version of rp
0
617
3,083,671,224
IssuesEvent
2015-08-24 10:28:51
Wikitalia/edgesense
https://api.github.com/repos/Wikitalia/edgesense
opened
Enable analysis to concentrate on last XX months of community
enhancement processing
It would be useful to have an analysis and visualization for the last period of time (year / six months), in addition to the all time network. It gives an idea on the status quo and who is active now. Also it is useful for older communities as InnovatoriPA were there are a lot of old nodes from the beginning which are not active any more and the shape of the network could be somehow different now with the only active ones. Also it is needed for activity management and in showing how each activity changed by time.
1.0
Enable analysis to concentrate on last XX months of community - It would be useful to have an analysis and visualization for the last period of time (year / six months), in addition to the all time network. It gives an idea on the status quo and who is active now. Also it is useful for older communities as InnovatoriPA were there are a lot of old nodes from the beginning which are not active any more and the shape of the network could be somehow different now with the only active ones. Also it is needed for activity management and in showing how each activity changed by time.
process
enable analysis to concentrate on last xx months of community it would be useful to have an analysis and visualization for the last period of time year six months in addition to the all time network it gives an idea on the status quo and who is active now also it is useful for older communities as innovatoripa were there are a lot of old nodes from the beginning which are not active any more and the shape of the network could be somehow different now with the only active ones also it is needed for activity management and in showing how each activity changed by time
1
36,828
8,148,946,130
IssuesEvent
2018-08-22 08:02:34
codl/forget
https://api.github.com/repos/codl/forget
opened
support new twitter archive format
defect
reported by `rrix@cybre.space` <https://cybre.space/@rrix/100591445673683791> > hey, it looks like the Twitter archive format changed at some point that makes it not work with Forget. there's no longer a data/js/tweets dir with monthly files, just a big jsonp file (67mib in my case) in the root of the zip
1.0
support new twitter archive format - reported by `rrix@cybre.space` <https://cybre.space/@rrix/100591445673683791> > hey, it looks like the Twitter archive format changed at some point that makes it not work with Forget. there's no longer a data/js/tweets dir with monthly files, just a big jsonp file (67mib in my case) in the root of the zip
non_process
support new twitter archive format reported by rrix cybre space hey it looks like the twitter archive format changed at some point that makes it not work with forget there s no longer a data js tweets dir with monthly files just a big jsonp file in my case in the root of the zip
0
122,876
16,372,799,056
IssuesEvent
2021-05-15 13:42:29
TCastus/mobilite2-front
https://api.github.com/repos/TCastus/mobilite2-front
closed
Design de la page d'accueil
design
Sur Canva, proposer une interface pour la page d'accueil avant l'implémentation
1.0
Design de la page d'accueil - Sur Canva, proposer une interface pour la page d'accueil avant l'implémentation
non_process
design de la page d accueil sur canva proposer une interface pour la page d accueil avant l implémentation
0