Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
75,619
| 3,470,203,071
|
IssuesEvent
|
2015-12-23 05:51:51
|
WebDevJL/EasyMvc
|
https://api.github.com/repos/WebDevJL/EasyMvc
|
closed
|
Generate any kind of class given the minimum necessary info
|
priority:high
|
Enable to specific:
- Class description => ok
- Class name =>ok
- Class derivation =>
- Class interface to implement =>
- Class destination folder (will be used to calculate the namespace) => ok
- Class properties with generation of get/set if necessary => later
- Class methods with name, parameters (using type-hinting) and default return type => later
NB: If return type is an object, the value provided will be the namespace of the Class to instanciate.
|
1.0
|
Generate any kind of class given the minimum necessary info - Enable to specific:
- Class description => ok
- Class name =>ok
- Class derivation =>
- Class interface to implement =>
- Class destination folder (will be used to calculate the namespace) => ok
- Class properties with generation of get/set if necessary => later
- Class methods with name, parameters (using type-hinting) and default return type => later
NB: If return type is an object, the value provided will be the namespace of the Class to instanciate.
|
non_process
|
generate any kind of class given the minimum necessary info enable to specific class description ok class name ok class derivation class interface to implement class destination folder will be used to calculate the namespace ok class properties with generation of get set if necessary later class methods with name parameters using type hinting and default return type later nb if return type is an object the value provided will be the namespace of the class to instanciate
| 0
|
9,283
| 12,303,889,802
|
IssuesEvent
|
2020-05-11 19:30:20
|
nextgenhealthcare/connect
|
https://api.github.com/repos/nextgenhealthcare/connect
|
closed
|
Allow users to return a Response from the Postprocessor script
|
post postprocessor processor response responseMap responsemap.put
|
Instead of making users call responseMap.put to place response map variables (and then having to select that variable in the source connector panel), it would be helpful to allow users to choose "Postprocessor" in the response combo box, and then simply return a response from the postprocessor. This would be the new Donkey Response object, so the ResponseFactory and reference list would need to be updated as well.
Imported Issue. Original Details:
Jira Issue Key: MIRTH-2279
Reporter: narupley
Created: 2012-11-28T13:52:00.000-0800
|
2.0
|
Allow users to return a Response from the Postprocessor script - Instead of making users call responseMap.put to place response map variables (and then having to select that variable in the source connector panel), it would be helpful to allow users to choose "Postprocessor" in the response combo box, and then simply return a response from the postprocessor. This would be the new Donkey Response object, so the ResponseFactory and reference list would need to be updated as well.
Imported Issue. Original Details:
Jira Issue Key: MIRTH-2279
Reporter: narupley
Created: 2012-11-28T13:52:00.000-0800
|
process
|
allow users to return a response from the postprocessor script instead of making users call responsemap put to place response map variables and then having to select that variable in the source connector panel it would be helpful to allow users to choose postprocessor in the response combo box and then simply return a response from the postprocessor this would be the new donkey response object so the responsefactory and reference list would need to be updated as well imported issue original details jira issue key mirth reporter narupley created
| 1
|
15,360
| 19,531,218,492
|
IssuesEvent
|
2021-12-30 17:11:39
|
pytorch/pytorch
|
https://api.github.com/repos/pytorch/pytorch
|
closed
|
multiprocessing ProcessException (and subclasses) can't be pickled/unpickled
|
module: multiprocessing triaged
|
### 🐛 Describe the bug
Error instances that inherit from [`ProcessException`](https://github.com/pytorch/pytorch/blob/1065739781cae67f7861beaaceadb736e5e52271/torch/multiprocessing/spawn.py#L12) cannot be picked/unpickled. For example:
```python
>>> import dill
>>> from torch.multiprocessing import ProcessRaisedException
>>> e = ProcessRaisedException("Oh no!", 1, 1)
>>> dill.loads(dill.dumps(e))
```
```
Traceback (most recent call last):
dill.loads(dill.dumps(e))
File ".../lib/python3.9/site-packages/dill/_dill.py", line 327, in loads
return load(file, ignore, **kwds)
File ".../lib/python3.9/site-packages/dill/_dill.py", line 313, in load
return Unpickler(file, ignore=ignore, **kwds).load()
File ".../lib/python3.9/site-packages/dill/_dill.py", line 525, in load
obj = StockUnpickler.load(self)
TypeError: __init__() missing 2 required positional arguments: 'error_index' and 'error_pid'
```
This is easy to fix and I wouldn't mind submitting a PR to do so. We'd just need to override `__reduce__` on `ProcessException`:
```python
class ProcessException(Exception):
__slots__ = ["error_index", "error_pid"]
def __init__(self, msg: str, error_index: int, pid: int):
super().__init__(msg)
self.error_index = error_index
self.pid = pid
def __reduce__(self):
return (self.__class__, self.args + (self.error_index, self.pid), {})
```
### Versions
Collecting environment information...
PyTorch version: 1.10.1+cu113
Is debug build: False
CUDA used to build PyTorch: 11.3
ROCM used to build PyTorch: N/A
OS: Ubuntu 20.04.3 LTS (x86_64)
GCC version: (Ubuntu 9.3.0-17ubuntu1~20.04) 9.3.0
Clang version: Could not collect
CMake version: version 3.16.3
Libc version: glibc-2.31
Python version: 3.9.7 (default, Sep 9 2021, 23:20:13) [GCC 9.3.0] (64-bit runtime)
Python platform: Linux-5.11.0-1023-gcp-x86_64-with-glibc2.31
Is CUDA available: True
CUDA runtime version: Could not collect
GPU models and configuration:
GPU 0: NVIDIA A100-SXM4-40GB
GPU 1: NVIDIA A100-SXM4-40GB
Nvidia driver version: 470.86
cuDNN version: Could not collect
HIP runtime version: N/A
MIOpen runtime version: N/A
Versions of relevant libraries:
[pip3] mypy==0.910
[pip3] mypy-extensions==0.4.3
[pip3] numpy==1.21.2
[pip3] torch==1.10.1+cu113
[conda] Could not collect
cc @VitalyFedyunin
|
1.0
|
multiprocessing ProcessException (and subclasses) can't be pickled/unpickled - ### 🐛 Describe the bug
Error instances that inherit from [`ProcessException`](https://github.com/pytorch/pytorch/blob/1065739781cae67f7861beaaceadb736e5e52271/torch/multiprocessing/spawn.py#L12) cannot be picked/unpickled. For example:
```python
>>> import dill
>>> from torch.multiprocessing import ProcessRaisedException
>>> e = ProcessRaisedException("Oh no!", 1, 1)
>>> dill.loads(dill.dumps(e))
```
```
Traceback (most recent call last):
dill.loads(dill.dumps(e))
File ".../lib/python3.9/site-packages/dill/_dill.py", line 327, in loads
return load(file, ignore, **kwds)
File ".../lib/python3.9/site-packages/dill/_dill.py", line 313, in load
return Unpickler(file, ignore=ignore, **kwds).load()
File ".../lib/python3.9/site-packages/dill/_dill.py", line 525, in load
obj = StockUnpickler.load(self)
TypeError: __init__() missing 2 required positional arguments: 'error_index' and 'error_pid'
```
This is easy to fix and I wouldn't mind submitting a PR to do so. We'd just need to override `__reduce__` on `ProcessException`:
```python
class ProcessException(Exception):
__slots__ = ["error_index", "error_pid"]
def __init__(self, msg: str, error_index: int, pid: int):
super().__init__(msg)
self.error_index = error_index
self.pid = pid
def __reduce__(self):
return (self.__class__, self.args + (self.error_index, self.pid), {})
```
### Versions
Collecting environment information...
PyTorch version: 1.10.1+cu113
Is debug build: False
CUDA used to build PyTorch: 11.3
ROCM used to build PyTorch: N/A
OS: Ubuntu 20.04.3 LTS (x86_64)
GCC version: (Ubuntu 9.3.0-17ubuntu1~20.04) 9.3.0
Clang version: Could not collect
CMake version: version 3.16.3
Libc version: glibc-2.31
Python version: 3.9.7 (default, Sep 9 2021, 23:20:13) [GCC 9.3.0] (64-bit runtime)
Python platform: Linux-5.11.0-1023-gcp-x86_64-with-glibc2.31
Is CUDA available: True
CUDA runtime version: Could not collect
GPU models and configuration:
GPU 0: NVIDIA A100-SXM4-40GB
GPU 1: NVIDIA A100-SXM4-40GB
Nvidia driver version: 470.86
cuDNN version: Could not collect
HIP runtime version: N/A
MIOpen runtime version: N/A
Versions of relevant libraries:
[pip3] mypy==0.910
[pip3] mypy-extensions==0.4.3
[pip3] numpy==1.21.2
[pip3] torch==1.10.1+cu113
[conda] Could not collect
cc @VitalyFedyunin
|
process
|
multiprocessing processexception and subclasses can t be pickled unpickled 🐛 describe the bug error instances that inherit from cannot be picked unpickled for example python import dill from torch multiprocessing import processraisedexception e processraisedexception oh no dill loads dill dumps e traceback most recent call last dill loads dill dumps e file lib site packages dill dill py line in loads return load file ignore kwds file lib site packages dill dill py line in load return unpickler file ignore ignore kwds load file lib site packages dill dill py line in load obj stockunpickler load self typeerror init missing required positional arguments error index and error pid this is easy to fix and i wouldn t mind submitting a pr to do so we d just need to override reduce on processexception python class processexception exception slots def init self msg str error index int pid int super init msg self error index error index self pid pid def reduce self return self class self args self error index self pid versions collecting environment information pytorch version is debug build false cuda used to build pytorch rocm used to build pytorch n a os ubuntu lts gcc version ubuntu clang version could not collect cmake version version libc version glibc python version default sep bit runtime python platform linux gcp with is cuda available true cuda runtime version could not collect gpu models and configuration gpu nvidia gpu nvidia nvidia driver version cudnn version could not collect hip runtime version n a miopen runtime version n a versions of relevant libraries mypy mypy extensions numpy torch could not collect cc vitalyfedyunin
| 1
|
15,651
| 19,846,744,120
|
IssuesEvent
|
2022-01-21 07:34:21
|
ooi-data/CE04OSSM-RID27-02-FLORTD000-recovered_host-flort_sample
|
https://api.github.com/repos/ooi-data/CE04OSSM-RID27-02-FLORTD000-recovered_host-flort_sample
|
opened
|
🛑 Processing failed: ValueError
|
process
|
## Overview
`ValueError` found in `processing_task` task during run ended on 2022-01-21T07:34:20.963174.
## Details
Flow name: `CE04OSSM-RID27-02-FLORTD000-recovered_host-flort_sample`
Task name: `processing_task`
Error type: `ValueError`
Error message: not enough values to unpack (expected 3, got 0)
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing
final_path = finalize_data_stream(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream
append_to_zarr(mod_ds, final_store, enc, logger=logger)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr
_append_zarr(store, mod_ds)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr
existing_arr.append(var_data.values)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values
return _as_array_or_item(self._data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item
data = np.asarray(data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__
x = self.compute()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute
(result,) = compute(self, traverse=False, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute
results = schedule(dsk, keys, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get
results = get_async(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async
raise_exception(exc, tb)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise
raise exc
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task
result = _execute_task(task, data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task
return func(*(_execute_task(a, cache) for a in args))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter
c = np.asarray(c)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__
self._ensure_cached()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached
self.array = NumpyIndexingAdapter(np.asarray(self.array))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__
return array[key.tuple]
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__
return self.get_basic_selection(selection, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection
return self._get_basic_selection_nd(selection=selection, out=out,
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd
return self._get_selection(indexer=indexer, out=out, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection
lchunk_coords, lchunk_selection, lout_selection = zip(*indexer)
ValueError: not enough values to unpack (expected 3, got 0)
```
</details>
|
1.0
|
🛑 Processing failed: ValueError - ## Overview
`ValueError` found in `processing_task` task during run ended on 2022-01-21T07:34:20.963174.
## Details
Flow name: `CE04OSSM-RID27-02-FLORTD000-recovered_host-flort_sample`
Task name: `processing_task`
Error type: `ValueError`
Error message: not enough values to unpack (expected 3, got 0)
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing
final_path = finalize_data_stream(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream
append_to_zarr(mod_ds, final_store, enc, logger=logger)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr
_append_zarr(store, mod_ds)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr
existing_arr.append(var_data.values)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values
return _as_array_or_item(self._data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item
data = np.asarray(data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__
x = self.compute()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute
(result,) = compute(self, traverse=False, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute
results = schedule(dsk, keys, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get
results = get_async(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async
raise_exception(exc, tb)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise
raise exc
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task
result = _execute_task(task, data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task
return func(*(_execute_task(a, cache) for a in args))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter
c = np.asarray(c)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__
self._ensure_cached()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached
self.array = NumpyIndexingAdapter(np.asarray(self.array))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__
return array[key.tuple]
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__
return self.get_basic_selection(selection, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection
return self._get_basic_selection_nd(selection=selection, out=out,
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd
return self._get_selection(indexer=indexer, out=out, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection
lchunk_coords, lchunk_selection, lout_selection = zip(*indexer)
ValueError: not enough values to unpack (expected 3, got 0)
```
</details>
|
process
|
🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name recovered host flort sample task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got
| 1
|
68,422
| 8,287,882,212
|
IssuesEvent
|
2018-09-19 10:09:34
|
cosmos/voyager
|
https://api.github.com/repos/cosmos/voyager
|
opened
|
Change PageValidator boxes background color
|
design staking1
|
Description:
<!-- Steps to reproduce, logs, and screenshots are helpful for us to resolve the bug -->
As per our discussion in #1317. The current boxes are similar to our input form fields. Let's change the background color to make it more distinguishable.
|
1.0
|
Change PageValidator boxes background color - Description:
<!-- Steps to reproduce, logs, and screenshots are helpful for us to resolve the bug -->
As per our discussion in #1317. The current boxes are similar to our input form fields. Let's change the background color to make it more distinguishable.
|
non_process
|
change pagevalidator boxes background color description as per our discussion in the current boxes are similar to our input form fields let s change the background color to make it more distinguishable
| 0
|
43,285
| 9,415,184,758
|
IssuesEvent
|
2019-04-10 12:03:08
|
junhoyeo/Dimicigan-Chrome
|
https://api.github.com/repos/junhoyeo/Dimicigan-Chrome
|
closed
|
Generic Object Injection Sink (security/detect-object-injection)
|
code-style
|
### [Codacy](https://app.codacy.com/app/junhoyeo/dimicigan-chrome/commit?cid=339205182) detected an issue:
#### Message: `Generic Object Injection Sink (security/detect-object-injection)`
#### Occurred on:
+ **Commit**: 3b6391c93a2660fcae6b75cb3c57fcad8c46eb6c
+ **File**: [src/pages/Index.vue](https://github.com/junhoyeo/dimicigan-chrome/blob/3b6391c93a2660fcae6b75cb3c57fcad8c46eb6c/src/pages/Index.vue)
+ **LineNum**: [80](https://github.com/junhoyeo/dimicigan-chrome/blob/3b6391c93a2660fcae6b75cb3c57fcad8c46eb6c/src/pages/Index.vue#L80)
+ **Code**: `subject: today[i],`
#### Currently on:
+ **Commit**: 6a2f84ade3db2d54eda829ca26cee9147d0407fb
+ **File**: [src/pages/Index.vue](https://github.com/junhoyeo/dimicigan-chrome/blob/6a2f84ade3db2d54eda829ca26cee9147d0407fb/src/pages/Index.vue)
+ **LineNum**: [80](https://github.com/junhoyeo/dimicigan-chrome/blob/6a2f84ade3db2d54eda829ca26cee9147d0407fb/src/pages/Index.vue#L80)
|
1.0
|
Generic Object Injection Sink (security/detect-object-injection) - ### [Codacy](https://app.codacy.com/app/junhoyeo/dimicigan-chrome/commit?cid=339205182) detected an issue:
#### Message: `Generic Object Injection Sink (security/detect-object-injection)`
#### Occurred on:
+ **Commit**: 3b6391c93a2660fcae6b75cb3c57fcad8c46eb6c
+ **File**: [src/pages/Index.vue](https://github.com/junhoyeo/dimicigan-chrome/blob/3b6391c93a2660fcae6b75cb3c57fcad8c46eb6c/src/pages/Index.vue)
+ **LineNum**: [80](https://github.com/junhoyeo/dimicigan-chrome/blob/3b6391c93a2660fcae6b75cb3c57fcad8c46eb6c/src/pages/Index.vue#L80)
+ **Code**: `subject: today[i],`
#### Currently on:
+ **Commit**: 6a2f84ade3db2d54eda829ca26cee9147d0407fb
+ **File**: [src/pages/Index.vue](https://github.com/junhoyeo/dimicigan-chrome/blob/6a2f84ade3db2d54eda829ca26cee9147d0407fb/src/pages/Index.vue)
+ **LineNum**: [80](https://github.com/junhoyeo/dimicigan-chrome/blob/6a2f84ade3db2d54eda829ca26cee9147d0407fb/src/pages/Index.vue#L80)
|
non_process
|
generic object injection sink security detect object injection detected an issue message generic object injection sink security detect object injection occurred on commit file linenum code subject today currently on commit file linenum
| 0
|
13,976
| 16,748,304,175
|
IssuesEvent
|
2021-06-11 18:38:32
|
sysflow-telemetry/sf-docs
|
https://api.github.com/repos/sysflow-telemetry/sf-docs
|
opened
|
Add a command-line flag for the processor to parse configuration and policy files as a standalone tool
|
enhancement sf-processor
|
**Indicate project**
Processor
**Describe the feature you'd like**
A command-line flag for the processor to parse configuration and policy files for syntax errors.
Example:
$> sfprocessor -log=info -config=pipeline.json `-test`
|
1.0
|
Add a command-line flag for the processor to parse configuration and policy files as a standalone tool - **Indicate project**
Processor
**Describe the feature you'd like**
A command-line flag for the processor to parse configuration and policy files for syntax errors.
Example:
$> sfprocessor -log=info -config=pipeline.json `-test`
|
process
|
add a command line flag for the processor to parse configuration and policy files as a standalone tool indicate project processor describe the feature you d like a command line flag for the processor to parse configuration and policy files for syntax errors example sfprocessor log info config pipeline json test
| 1
|
260,529
| 27,784,521,620
|
IssuesEvent
|
2023-03-17 01:14:44
|
DavidSpek/kubeflow
|
https://api.github.com/repos/DavidSpek/kubeflow
|
opened
|
CVE-2023-28155 (Medium) detected in request-2.88.0.tgz, request-2.88.2.tgz
|
Mend: dependency security vulnerability
|
## CVE-2023-28155 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>request-2.88.0.tgz</b>, <b>request-2.88.2.tgz</b></p></summary>
<p>
<details><summary><b>request-2.88.0.tgz</b></p></summary>
<p>Simplified HTTP request client.</p>
<p>Library home page: <a href="https://registry.npmjs.org/request/-/request-2.88.0.tgz">https://registry.npmjs.org/request/-/request-2.88.0.tgz</a></p>
<p>Path to dependency file: /components/crud-web-apps/volumes/frontend/package.json</p>
<p>Path to vulnerable library: /components/crud-web-apps/volumes/frontend/node_modules/request/package.json,/components/centraldashboard/node_modules/request/package.json</p>
<p>
Dependency Hierarchy:
- client-node-0.12.3.tgz (Root Library)
- :x: **request-2.88.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>request-2.88.2.tgz</b></p></summary>
<p>Simplified HTTP request client.</p>
<p>Library home page: <a href="https://registry.npmjs.org/request/-/request-2.88.2.tgz">https://registry.npmjs.org/request/-/request-2.88.2.tgz</a></p>
<p>Path to dependency file: /components/crud-web-apps/common/frontend/kubeflow-common-lib/package.json</p>
<p>Path to vulnerable library: /components/crud-web-apps/common/frontend/kubeflow-common-lib/node_modules/request/package.json,/components/crud-web-apps/tensorboards/frontend/node_modules/request/package.json,/components/crud-web-apps/jupyter/frontend/node_modules/request/package.json,/components/crud-web-apps/volumes/frontend/node_modules/node-gyp/node_modules/request/package.json</p>
<p>
Dependency Hierarchy:
- client-node-0.12.2.tgz (Root Library)
- :x: **request-2.88.2.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
** UNSUPPORTED WHEN ASSIGNED ** The Request package through 2.88.1 for Node.js allows a bypass of SSRF mitigations via an attacker-controller server that does a cross-protocol redirect (HTTP to HTTPS, or HTTPS to HTTP). NOTE: This vulnerability only affects products that are no longer supported by the maintainer.
<p>Publish Date: 2023-03-16
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-28155>CVE-2023-28155</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2023-28155 (Medium) detected in request-2.88.0.tgz, request-2.88.2.tgz - ## CVE-2023-28155 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>request-2.88.0.tgz</b>, <b>request-2.88.2.tgz</b></p></summary>
<p>
<details><summary><b>request-2.88.0.tgz</b></p></summary>
<p>Simplified HTTP request client.</p>
<p>Library home page: <a href="https://registry.npmjs.org/request/-/request-2.88.0.tgz">https://registry.npmjs.org/request/-/request-2.88.0.tgz</a></p>
<p>Path to dependency file: /components/crud-web-apps/volumes/frontend/package.json</p>
<p>Path to vulnerable library: /components/crud-web-apps/volumes/frontend/node_modules/request/package.json,/components/centraldashboard/node_modules/request/package.json</p>
<p>
Dependency Hierarchy:
- client-node-0.12.3.tgz (Root Library)
- :x: **request-2.88.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>request-2.88.2.tgz</b></p></summary>
<p>Simplified HTTP request client.</p>
<p>Library home page: <a href="https://registry.npmjs.org/request/-/request-2.88.2.tgz">https://registry.npmjs.org/request/-/request-2.88.2.tgz</a></p>
<p>Path to dependency file: /components/crud-web-apps/common/frontend/kubeflow-common-lib/package.json</p>
<p>Path to vulnerable library: /components/crud-web-apps/common/frontend/kubeflow-common-lib/node_modules/request/package.json,/components/crud-web-apps/tensorboards/frontend/node_modules/request/package.json,/components/crud-web-apps/jupyter/frontend/node_modules/request/package.json,/components/crud-web-apps/volumes/frontend/node_modules/node-gyp/node_modules/request/package.json</p>
<p>
Dependency Hierarchy:
- client-node-0.12.2.tgz (Root Library)
- :x: **request-2.88.2.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
** UNSUPPORTED WHEN ASSIGNED ** The Request package through 2.88.1 for Node.js allows a bypass of SSRF mitigations via an attacker-controller server that does a cross-protocol redirect (HTTP to HTTPS, or HTTPS to HTTP). NOTE: This vulnerability only affects products that are no longer supported by the maintainer.
<p>Publish Date: 2023-03-16
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-28155>CVE-2023-28155</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in request tgz request tgz cve medium severity vulnerability vulnerable libraries request tgz request tgz request tgz simplified http request client library home page a href path to dependency file components crud web apps volumes frontend package json path to vulnerable library components crud web apps volumes frontend node modules request package json components centraldashboard node modules request package json dependency hierarchy client node tgz root library x request tgz vulnerable library request tgz simplified http request client library home page a href path to dependency file components crud web apps common frontend kubeflow common lib package json path to vulnerable library components crud web apps common frontend kubeflow common lib node modules request package json components crud web apps tensorboards frontend node modules request package json components crud web apps jupyter frontend node modules request package json components crud web apps volumes frontend node modules node gyp node modules request package json dependency hierarchy client node tgz root library x request tgz vulnerable library found in base branch master vulnerability details unsupported when assigned the request package through for node js allows a bypass of ssrf mitigations via an attacker controller server that does a cross protocol redirect http to https or https to http note this vulnerability only affects products that are no longer supported by the maintainer publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with mend
| 0
|
5,590
| 3,251,785,631
|
IssuesEvent
|
2015-10-19 11:52:14
|
robertocarroll/ideas
|
https://api.github.com/repos/robertocarroll/ideas
|
opened
|
Stories around photos
|
code writing
|
Writing around photos, Barthes, see also http://www.thehypertext.com/2015/04/11/word-camera/
1. Found photos from parents' album
2. Process them with Clarifai API
3. Use tags as inspiration to write - short things -
4. Put the writing back with the images
|
1.0
|
Stories around photos - Writing around photos, Barthes, see also http://www.thehypertext.com/2015/04/11/word-camera/
1. Found photos from parents' album
2. Process them with Clarifai API
3. Use tags as inspiration to write - short things -
4. Put the writing back with the images
|
non_process
|
stories around photos writing around photos barthes see also found photos from parents album process them with clarifai api use tags as inspiration to write short things put the writing back with the images
| 0
|
277,635
| 24,090,807,673
|
IssuesEvent
|
2022-09-19 14:37:17
|
eclipse-openj9/openj9
|
https://api.github.com/repos/eclipse-openj9/openj9
|
opened
|
jdk19 MiniMix_aot_5m_0 hang
|
test failure
| ERROR: type should be string, got "https://openj9-jenkins.osuosl.org/job/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/8\r\nMiniMix_aot_5m_0\r\n\r\nThere is a javacore and core created after the hang, but also another core.1202972 which isn't remained. Likely caused by sending signals since it's create a minute after the renamed core file.\r\n\r\nhttps://openj9-artifactory.osuosl.org/artifactory/ci-openj9/Test/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/8/system_test_output.tar.gz\r\n\r\n```\r\nLT 06:01:05.015 - Completed 20.0%. Number of tests started=10571 (+4241)\r\nSTF 06:05:03.037 - Heartbeat: Process LT is still running\r\nSTF 06:10:03.284 - Heartbeat: Process LT is still running\r\nSTF 06:15:03.166 - Heartbeat: Process LT is still running\r\nSTF 06:20:03.478 - Heartbeat: Process LT is still running\r\nSTF 06:25:03.249 - Heartbeat: Process LT is still running\r\nSTF 06:30:03.295 - Heartbeat: Process LT is still running\r\nSTF 06:35:03.021 - Heartbeat: Process LT is still running\r\nSTF 06:40:03.236 - Heartbeat: Process LT is still running\r\nSTF 06:45:03.344 - Heartbeat: Process LT is still running\r\nSTF 06:50:03.261 - Heartbeat: Process LT is still running\r\nSTF 06:55:03.194 - Heartbeat: Process LT is still running\r\nSTF 07:00:03.453 - Heartbeat: Process LT is still running\r\nSTF 07:05:03.250 - Heartbeat: Process LT is still running\r\nSTF 07:05:04.253 - **FAILED** Process LT has timed out\r\nSTF 07:05:04.254 - Collecting dumps for: LT \r\nSTF 07:05:04.254 - Sending SIG 3 to the java process to generate a javacore\r\nSTF 07:05:04.255 - Running command: kill -3 1202972\r\nSTF 07:05:04.255 - Redirecting stderr to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_3.stderr\r\nSTF 07:05:04.255 - Redirecting stdout to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_3.stdout\r\nSTF 07:05:04.256 - Pausing for 30 seconds\r\nSTF 07:05:34.259 - Sending SIG 3 to the java process to generate a javacore\r\nSTF 07:05:34.259 - Running command: kill -3 1202972\r\nSTF 07:05:34.259 - Redirecting stderr to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_3.stderr\r\nSTF 07:05:34.259 - Redirecting stdout to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_3.stdout\r\nSTF 07:05:34.260 - Pausing for 30 seconds\r\nSTF 07:06:04.262 - Sending SIG 3 to the java process to generate a javacore\r\nSTF 07:06:04.310 - Running command: kill -3 1202972\r\nSTF 07:06:04.310 - Redirecting stderr to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_3.stderr\r\nSTF 07:06:04.310 - Redirecting stdout to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_3.stdout\r\nSTF 07:06:04.311 - Pausing for 30 seconds\r\nSTF 07:06:34.311 - Sending SIGABRT (kill -6) to the java process to generate a core\r\nSTF 07:06:34.312 - Running command: kill -6 1202972\r\nSTF 07:06:34.312 - Redirecting stderr to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_6.stderr\r\nSTF 07:06:34.312 - Redirecting stdout to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_6.stdout\r\nSTF 07:06:34.313 - Pausing for 30 seconds\r\nSTF 07:07:04.315 - Sending SIGXCPU (kill -24) to the java process to generate an OS dump\r\nSTF 07:07:04.315 - Running command: kill -24 1202972\r\nSTF 07:07:04.315 - Redirecting stderr to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_24.stderr\r\nSTF 07:07:04.315 - Redirecting stdout to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_24.stdout\r\nLT stderr JVMDUMP039I Processing dump event \"user\", detail \"\" at 2022/09/17 07:05:04 - please wait.\r\nLT stderr JVMDUMP032I JVM requested System dump using '/home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/core.20220917.070504.1202972.0001.dmp' in response to an event\r\nLT stderr JVMDUMP010I System dump written to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/core.20220917.070504.1202972.0001.dmp\r\nLT stderr JVMDUMP032I JVM requested Java dump using '/home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/javacore.20220917.070504.1202972.0002.txt' in response to an event\r\nLT stderr JVMDUMP010I Java dump written to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/javacore.20220917.070504.1202972.0002.txt\r\nLT stderr JVMDUMP013I Processed dump event \"user\", detail \"\".\r\n```"
|
1.0
|
jdk19 MiniMix_aot_5m_0 hang - https://openj9-jenkins.osuosl.org/job/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/8
MiniMix_aot_5m_0
There is a javacore and core created after the hang, but also another core.1202972 which isn't remained. Likely caused by sending signals since it's create a minute after the renamed core file.
https://openj9-artifactory.osuosl.org/artifactory/ci-openj9/Test/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/8/system_test_output.tar.gz
```
LT 06:01:05.015 - Completed 20.0%. Number of tests started=10571 (+4241)
STF 06:05:03.037 - Heartbeat: Process LT is still running
STF 06:10:03.284 - Heartbeat: Process LT is still running
STF 06:15:03.166 - Heartbeat: Process LT is still running
STF 06:20:03.478 - Heartbeat: Process LT is still running
STF 06:25:03.249 - Heartbeat: Process LT is still running
STF 06:30:03.295 - Heartbeat: Process LT is still running
STF 06:35:03.021 - Heartbeat: Process LT is still running
STF 06:40:03.236 - Heartbeat: Process LT is still running
STF 06:45:03.344 - Heartbeat: Process LT is still running
STF 06:50:03.261 - Heartbeat: Process LT is still running
STF 06:55:03.194 - Heartbeat: Process LT is still running
STF 07:00:03.453 - Heartbeat: Process LT is still running
STF 07:05:03.250 - Heartbeat: Process LT is still running
STF 07:05:04.253 - **FAILED** Process LT has timed out
STF 07:05:04.254 - Collecting dumps for: LT
STF 07:05:04.254 - Sending SIG 3 to the java process to generate a javacore
STF 07:05:04.255 - Running command: kill -3 1202972
STF 07:05:04.255 - Redirecting stderr to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_3.stderr
STF 07:05:04.255 - Redirecting stdout to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_3.stdout
STF 07:05:04.256 - Pausing for 30 seconds
STF 07:05:34.259 - Sending SIG 3 to the java process to generate a javacore
STF 07:05:34.259 - Running command: kill -3 1202972
STF 07:05:34.259 - Redirecting stderr to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_3.stderr
STF 07:05:34.259 - Redirecting stdout to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_3.stdout
STF 07:05:34.260 - Pausing for 30 seconds
STF 07:06:04.262 - Sending SIG 3 to the java process to generate a javacore
STF 07:06:04.310 - Running command: kill -3 1202972
STF 07:06:04.310 - Redirecting stderr to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_3.stderr
STF 07:06:04.310 - Redirecting stdout to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_3.stdout
STF 07:06:04.311 - Pausing for 30 seconds
STF 07:06:34.311 - Sending SIGABRT (kill -6) to the java process to generate a core
STF 07:06:34.312 - Running command: kill -6 1202972
STF 07:06:34.312 - Redirecting stderr to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_6.stderr
STF 07:06:34.312 - Redirecting stdout to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_6.stdout
STF 07:06:34.313 - Pausing for 30 seconds
STF 07:07:04.315 - Sending SIGXCPU (kill -24) to the java process to generate an OS dump
STF 07:07:04.315 - Running command: kill -24 1202972
STF 07:07:04.315 - Redirecting stderr to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_24.stderr
STF 07:07:04.315 - Redirecting stdout to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/1.LT.kill_24.stdout
LT stderr JVMDUMP039I Processing dump event "user", detail "" at 2022/09/17 07:05:04 - please wait.
LT stderr JVMDUMP032I JVM requested System dump using '/home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/core.20220917.070504.1202972.0001.dmp' in response to an event
LT stderr JVMDUMP010I System dump written to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/core.20220917.070504.1202972.0001.dmp
LT stderr JVMDUMP032I JVM requested Java dump using '/home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/javacore.20220917.070504.1202972.0002.txt' in response to an event
LT stderr JVMDUMP010I Java dump written to /home/jenkins/workspace/Test_openjdk19_j9_extended.system_s390x_linux_Nightly_testList_1/aqa-tests/TKG/output_1663392324450/MiniMix_aot_5m_0/20220917-060001-MixedLoadTest/results/javacore.20220917.070504.1202972.0002.txt
LT stderr JVMDUMP013I Processed dump event "user", detail "".
```
|
non_process
|
minimix aot hang minimix aot there is a javacore and core created after the hang but also another core which isn t remained likely caused by sending signals since it s create a minute after the renamed core file lt completed number of tests started stf heartbeat process lt is still running stf heartbeat process lt is still running stf heartbeat process lt is still running stf heartbeat process lt is still running stf heartbeat process lt is still running stf heartbeat process lt is still running stf heartbeat process lt is still running stf heartbeat process lt is still running stf heartbeat process lt is still running stf heartbeat process lt is still running stf heartbeat process lt is still running stf heartbeat process lt is still running stf heartbeat process lt is still running stf failed process lt has timed out stf collecting dumps for lt stf sending sig to the java process to generate a javacore stf running command kill stf redirecting stderr to home jenkins workspace test extended system linux nightly testlist aqa tests tkg output minimix aot mixedloadtest results lt kill stderr stf redirecting stdout to home jenkins workspace test extended system linux nightly testlist aqa tests tkg output minimix aot mixedloadtest results lt kill stdout stf pausing for seconds stf sending sig to the java process to generate a javacore stf running command kill stf redirecting stderr to home jenkins workspace test extended system linux nightly testlist aqa tests tkg output minimix aot mixedloadtest results lt kill stderr stf redirecting stdout to home jenkins workspace test extended system linux nightly testlist aqa tests tkg output minimix aot mixedloadtest results lt kill stdout stf pausing for seconds stf sending sig to the java process to generate a javacore stf running command kill stf redirecting stderr to home jenkins workspace test extended system linux nightly testlist aqa tests tkg output minimix aot mixedloadtest results lt kill stderr stf redirecting stdout to home jenkins workspace test extended system linux nightly testlist aqa tests tkg output minimix aot mixedloadtest results lt kill stdout stf pausing for seconds stf sending sigabrt kill to the java process to generate a core stf running command kill stf redirecting stderr to home jenkins workspace test extended system linux nightly testlist aqa tests tkg output minimix aot mixedloadtest results lt kill stderr stf redirecting stdout to home jenkins workspace test extended system linux nightly testlist aqa tests tkg output minimix aot mixedloadtest results lt kill stdout stf pausing for seconds stf sending sigxcpu kill to the java process to generate an os dump stf running command kill stf redirecting stderr to home jenkins workspace test extended system linux nightly testlist aqa tests tkg output minimix aot mixedloadtest results lt kill stderr stf redirecting stdout to home jenkins workspace test extended system linux nightly testlist aqa tests tkg output minimix aot mixedloadtest results lt kill stdout lt stderr processing dump event user detail at please wait lt stderr jvm requested system dump using home jenkins workspace test extended system linux nightly testlist aqa tests tkg output minimix aot mixedloadtest results core dmp in response to an event lt stderr system dump written to home jenkins workspace test extended system linux nightly testlist aqa tests tkg output minimix aot mixedloadtest results core dmp lt stderr jvm requested java dump using home jenkins workspace test extended system linux nightly testlist aqa tests tkg output minimix aot mixedloadtest results javacore txt in response to an event lt stderr java dump written to home jenkins workspace test extended system linux nightly testlist aqa tests tkg output minimix aot mixedloadtest results javacore txt lt stderr processed dump event user detail
| 0
|
21,169
| 28,140,404,000
|
IssuesEvent
|
2023-04-01 21:59:33
|
kserve/kserve
|
https://api.github.com/repos/kserve/kserve
|
closed
|
Parametrized container builds
|
kind/feature kserve/release-process
|
/kind feature
**Describe the solution you'd like**
Today, the base images are hardcoded in Dockerfiles, which makes it hard to rebuild the KServe components on custom images.
Parametrizing the Dockerfile base images would enable KServe providers to swap base images to improve security or integration with the provider's platform.
|
1.0
|
Parametrized container builds - /kind feature
**Describe the solution you'd like**
Today, the base images are hardcoded in Dockerfiles, which makes it hard to rebuild the KServe components on custom images.
Parametrizing the Dockerfile base images would enable KServe providers to swap base images to improve security or integration with the provider's platform.
|
process
|
parametrized container builds kind feature describe the solution you d like today the base images are hardcoded in dockerfiles which makes it hard to rebuild the kserve components on custom images parametrizing the dockerfile base images would enable kserve providers to swap base images to improve security or integration with the provider s platform
| 1
|
1,198
| 3,697,440,741
|
IssuesEvent
|
2016-02-27 17:39:39
|
pelias/fuzzy-tester
|
https://api.github.com/repos/pelias/fuzzy-tester
|
closed
|
Use Lat/Lon as one of the factors for fuzzy scoring
|
processed
|
One of the things that we care a lot about is "is this the right location". Right now, our testsuite doesn't take the location of the place into account, but instead relies on matching the labels of documents it knows to be correct.
But often there are several entries for a single place, an artifact of us importing from multiple data sources and only checking for duplication within the dataset. These documents are of varying quality, but all that matters to the user is "Is the place right".
In addition to these criteria, we should incorporate comparisons of distance from the "correct" document as well.
One possibility of how to consider this is to calculate distance from the "correct" target document for each query and have a gradient for scoring [e.g 0-10m, 10-30m, 30-100m, 500m-2k, 2k+] where the closer it is to the target the higher the score.
|
1.0
|
Use Lat/Lon as one of the factors for fuzzy scoring - One of the things that we care a lot about is "is this the right location". Right now, our testsuite doesn't take the location of the place into account, but instead relies on matching the labels of documents it knows to be correct.
But often there are several entries for a single place, an artifact of us importing from multiple data sources and only checking for duplication within the dataset. These documents are of varying quality, but all that matters to the user is "Is the place right".
In addition to these criteria, we should incorporate comparisons of distance from the "correct" document as well.
One possibility of how to consider this is to calculate distance from the "correct" target document for each query and have a gradient for scoring [e.g 0-10m, 10-30m, 30-100m, 500m-2k, 2k+] where the closer it is to the target the higher the score.
|
process
|
use lat lon as one of the factors for fuzzy scoring one of the things that we care a lot about is is this the right location right now our testsuite doesn t take the location of the place into account but instead relies on matching the labels of documents it knows to be correct but often there are several entries for a single place an artifact of us importing from multiple data sources and only checking for duplication within the dataset these documents are of varying quality but all that matters to the user is is the place right in addition to these criteria we should incorporate comparisons of distance from the correct document as well one possibility of how to consider this is to calculate distance from the correct target document for each query and have a gradient for scoring where the closer it is to the target the higher the score
| 1
|
12,258
| 14,787,266,031
|
IssuesEvent
|
2021-01-12 07:18:08
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[Mobile Apps] Studies list > Progress bar is not updated for study completion % in studies list
|
Bug P1 Process: Dev Process: Reopened Process: Tested dev
|
**Steps:**
1. Login to mobile
2. Enroll into a study
3. Complete some activities
4. Navigate to studies list
5. Observe the progress bar
**Actual**: Progress bar is not updated for study completion % in studies list - iOS
Progress bar is not updated for study completion % in studies list after logout and login - Android
**Expected**: Progress bar is should be updated for study completion % in studies list

Completion % updating in Dashboard:

|
3.0
|
[Mobile Apps] Studies list > Progress bar is not updated for study completion % in studies list - **Steps:**
1. Login to mobile
2. Enroll into a study
3. Complete some activities
4. Navigate to studies list
5. Observe the progress bar
**Actual**: Progress bar is not updated for study completion % in studies list - iOS
Progress bar is not updated for study completion % in studies list after logout and login - Android
**Expected**: Progress bar is should be updated for study completion % in studies list

Completion % updating in Dashboard:

|
process
|
studies list progress bar is not updated for study completion in studies list steps login to mobile enroll into a study complete some activities navigate to studies list observe the progress bar actual progress bar is not updated for study completion in studies list ios progress bar is not updated for study completion in studies list after logout and login android expected progress bar is should be updated for study completion in studies list completion updating in dashboard
| 1
|
164,701
| 6,254,181,696
|
IssuesEvent
|
2017-07-14 00:56:15
|
HabitRPG/habitica
|
https://api.github.com/repos/HabitRPG/habitica
|
closed
|
Mod Tools Desired
|
priority: important status: issue: suggestion-discussion
|
Figured we should open a ticket to start discussing some Moderator Tools that would make our lives easier.
Some things that would be helpful to me:
- Ability to edit/delete Challenges
- Possibly, ability to edit messages instead of just deleting them (maybe to put trigger warnings at the beginning, etc).
- Profanity filters. I know that there has been some debate about profanity filters, but it gets pretty tedious to filter out messages in the Tavern manually - and many of them are innocent mistakes. Even a popup with our stock message and a swapped in suggestion word (Curses! Pain in the neck! Dratitude!) would go a long way. Of course we'd occasionally get people going out of their way to circumnavigate it, but they're probably trolling anyway, so it would still minimize the hassle we have.
- Similarly, we should have something in place to prevent users from choosing usernames that contain language that isn't acceptable in the public chat rooms. I just received an angry email from someone who chose an inappropriate username unwittingly and when asked to correct it, got very upset. She had a good point, which is that if it is inappropriate, she shouldn't have been allowed to choose it. (She also had some... less helpful.... suggestions about what we should go do, but I'm choosing to extract the constructive part of the feedback.)
I'm sure there are plenty of other useful tools - these are just some that I've been wanting a lot recently. Other mods should definitely chime in with what they want!
@Alys @deilann @DanielTheBard @veeeeeee @lefnire @SabreCat @paglias
## <bountysource-plugin>
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/5197741-mod-tools-desired?utm_campaign=plugin&utm_content=tracker%2F68393&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F68393&utm_medium=issues&utm_source=github).
</bountysource-plugin>
|
1.0
|
Mod Tools Desired - Figured we should open a ticket to start discussing some Moderator Tools that would make our lives easier.
Some things that would be helpful to me:
- Ability to edit/delete Challenges
- Possibly, ability to edit messages instead of just deleting them (maybe to put trigger warnings at the beginning, etc).
- Profanity filters. I know that there has been some debate about profanity filters, but it gets pretty tedious to filter out messages in the Tavern manually - and many of them are innocent mistakes. Even a popup with our stock message and a swapped in suggestion word (Curses! Pain in the neck! Dratitude!) would go a long way. Of course we'd occasionally get people going out of their way to circumnavigate it, but they're probably trolling anyway, so it would still minimize the hassle we have.
- Similarly, we should have something in place to prevent users from choosing usernames that contain language that isn't acceptable in the public chat rooms. I just received an angry email from someone who chose an inappropriate username unwittingly and when asked to correct it, got very upset. She had a good point, which is that if it is inappropriate, she shouldn't have been allowed to choose it. (She also had some... less helpful.... suggestions about what we should go do, but I'm choosing to extract the constructive part of the feedback.)
I'm sure there are plenty of other useful tools - these are just some that I've been wanting a lot recently. Other mods should definitely chime in with what they want!
@Alys @deilann @DanielTheBard @veeeeeee @lefnire @SabreCat @paglias
## <bountysource-plugin>
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/5197741-mod-tools-desired?utm_campaign=plugin&utm_content=tracker%2F68393&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F68393&utm_medium=issues&utm_source=github).
</bountysource-plugin>
|
non_process
|
mod tools desired figured we should open a ticket to start discussing some moderator tools that would make our lives easier some things that would be helpful to me ability to edit delete challenges possibly ability to edit messages instead of just deleting them maybe to put trigger warnings at the beginning etc profanity filters i know that there has been some debate about profanity filters but it gets pretty tedious to filter out messages in the tavern manually and many of them are innocent mistakes even a popup with our stock message and a swapped in suggestion word curses pain in the neck dratitude would go a long way of course we d occasionally get people going out of their way to circumnavigate it but they re probably trolling anyway so it would still minimize the hassle we have similarly we should have something in place to prevent users from choosing usernames that contain language that isn t acceptable in the public chat rooms i just received an angry email from someone who chose an inappropriate username unwittingly and when asked to correct it got very upset she had a good point which is that if it is inappropriate she shouldn t have been allowed to choose it she also had some less helpful suggestions about what we should go do but i m choosing to extract the constructive part of the feedback i m sure there are plenty of other useful tools these are just some that i ve been wanting a lot recently other mods should definitely chime in with what they want alys deilann danielthebard veeeeeee lefnire sabrecat paglias want to back this issue we accept bounties via
| 0
|
13,461
| 15,946,025,585
|
IssuesEvent
|
2021-04-15 00:03:51
|
googleapis/release-please
|
https://api.github.com/repos/googleapis/release-please
|
closed
|
add tests for each language's releaser
|
type: process
|
we've split out the release logic into specific languages, e.g., ruby-yoshi, node, let's actually add some tests for each language.
|
1.0
|
add tests for each language's releaser - we've split out the release logic into specific languages, e.g., ruby-yoshi, node, let's actually add some tests for each language.
|
process
|
add tests for each language s releaser we ve split out the release logic into specific languages e g ruby yoshi node let s actually add some tests for each language
| 1
|
331,555
| 24,312,567,958
|
IssuesEvent
|
2022-09-30 00:56:38
|
roots/bud
|
https://api.github.com/repos/roots/bud
|
closed
|
[bug] bud-sass loader not registering, ignores files using a transpiler source
|
documentation
|
### Agreement
- [X] This is not a duplicate of an existing issue
- [X] I have read the [guidelines for Contributing to Roots Projects](https://github.com/roots/.github/blob/master/CONTRIBUTING.md)
- [X] This is not a personal support request that should be posted on the [Roots Discourse](https://discourse.roots.io/) community
### Describe the issue
I have a need to load styles/js from other directories than `./resources`. I wish to compile .scss files in a directory to their own CSS files.
I have created a test directory in the root of my sage theme, containing an empty `script.js` and an empty `style.scss`
I then add the test folder to the entry point:
```js
app
/**
* Application entrypoints
*/
.entry({
app: ["@scripts/app", "@styles/app"],
editor: ["@scripts/editor", "@styles/editor"],
test: [app.path("./test/script.js"), app.path("./test/style.scss")]
})
```
After that, I add a transpiler source to the js and sass rules:
```js
app.build.rules.js.setInclude([
bud => bud.path('@src'),
bud => bud.path('./test'),
])
```
This works fine, but the .scss file is ignored. So I go ahead and try add another source to the sass rule:
```js
app.build.rules.sass.setInclude([
bud => bud.path('@src'),
bud => bud.path('./test'),
])
```
Which throws an error:
` Type Error: Cannot read properties of undefined (reading 'setInclude') `
In my bud config, I console log out the `app.build.rules` and see there is no sass rule:
```js
{
yml: Rule {
_app: [Function (anonymous)],
include: [ [Function (anonymous)] ],
test: [Function (anonymous)],
use: [ 'yml' ]
},
webp: Rule {
_app: [Function (anonymous)],
test: [Function (anonymous)],
include: [ [Function (anonymous)] ],
type: 'asset/resource',
generator: [Function (anonymous)]
},
svg: Rule {
_app: [Function (anonymous)],
test: [Function (anonymous)],
include: [ [Function (anonymous)] ],
type: 'asset/resource',
generator: [Function: bound svgGenerator]
},
json: Rule {
_app: [Function (anonymous)],
type: 'json',
include: [ [Function (anonymous)] ],
test: [Function (anonymous)],
parser: [Function (anonymous)]
},
js: Rule {
_app: [Function (anonymous)],
test: [Function (anonymous)],
include: [ [Function (anonymous)] ],
use: [ [Item] ]
},
image: Rule {
_app: [Function (anonymous)],
test: [Function (anonymous)],
include: [ [Function (anonymous)] ],
type: 'asset/resource',
generator: [Function (anonymous)]
},
html: Rule {
_app: [Function (anonymous)],
include: [ [Function (anonymous)] ],
test: [Function (anonymous)],
use: [ 'html' ]
},
font: Rule {
_app: [Function (anonymous)],
type: 'asset',
test: [Function (anonymous)],
include: [ [Function (anonymous)] ],
generator: [Function (anonymous)]
},
cssModule: Rule {
_app: [Function (anonymous)],
test: [Function (anonymous)],
include: [ [Function (anonymous)] ],
use: [ 'precss', 'cssModule', 'postcss' ]
},
css: Rule {
_app: [Function (anonymous)],
test: [Function (anonymous)],
include: [ [Function (anonymous)] ],
use: [ 'precss', 'css', 'postcss' ]
}
}
```
### Expected Behavior
.scss files in additional entry points, that have a transpiler source - to be parsed using bud-sass
### Actual Behavior
This error is thrown: `Type Error: Cannot read properties of undefined (reading 'setInclude')`
### Steps To Reproduce
1. Create a test folder in the root of the teme
2. Add a blank script.js and style.scss to the folder
3. Add the entry points for the test folder into the bud config
4. Add transpiler sources for the js/sass rules
### version
6.4.4
### What package manager are you using?
npm
### version
8.15.0
### Logs
```zsh
Type Error: Cannot read properties of undefined (reading 'setInclude')
at default (file:///Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/bud.config.mjs:19:24)
at Configuration.run (file:///Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/node_modules/@roots/bud-framework/lib/configuration/configuration.js:30:26)
at file:///Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/node_modules/@roots/bud-framework/lib/configuration/index.js:14:29
at Array.map (<anonymous>)
at Module.process (file:///Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/node_modules/@roots/bud-framework/lib/configuration/index.js:13:60)
at Bud.run (file:///Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/node_modules/@roots/bud-framework/lib/methods/run.js:5:29)
at BuildDevelopmentCommand.runCommand (file:///Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/node_modules/@roots/bud/lib/cli/commands/build.base.js:192:24)
at BuildDevelopmentCommand.execute (file:///Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/node_modules/@roots/bud/lib/cli/commands/base.js:179:24)
at async BuildDevelopmentCommand.validateAndExecute (/Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/node_modules/clipanion/lib/advanced/Command.js:73:26)
at async Cli.run (/Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/node_modules/clipanion/lib/advanced/Cli.js:222:24)
```
### Configuration
```zsh
// @ts-check
/**
* Build configuration
*
* @see {@link https://bud.js.org/guides/configure}
* @param {import('@roots/bud').Bud} app
*/
export default async (app) => {
console.log(app.build.rules)
app.build.rules.js.setInclude([
bud => bud.path('@src'),
bud => bud.path('./test'),
])
app.build.rules.scss.setInclude([
bud => bud.path('@src'),
bud => bud.path('./test'),
])
app
/**
* Application entrypoints
*/
.entry({
app: ["@scripts/app", "@styles/app"],
editor: ["@scripts/editor", "@styles/editor"],
test: [app.path("./test/script.js"), app.path("./test/style.scss")]
})
/**
* Directory contents to be included in the compilation
*/
.assets(["images"])
/**
* Matched files trigger a page reload when modified
*/
.watch(["resources/views/**/*", "app/**/*"])
/**
* Proxy origin (`WP_HOME`)
*/
.proxy("http://example.test")
/**
* Development origin
*/
.serve("http://0.0.0.0:3000")
/**
* URI of the `public` directory
*/
.setPublicPath("/app/themes/sage/public/")
/**
* Generate WordPress `theme.json`
*
* @note This overwrites `theme.json` on every build.
*/
.wpjson
.settings({
color: {
custom: false,
customGradient: false,
defaultPalette: false,
defaultGradients: false,
},
custom: {
spacing: {},
typography: {
'font-size': {},
'line-height': {},
},
},
spacing: {
padding: true,
units: ['px', '%', 'em', 'rem', 'vw', 'vh'],
},
typography: {
customFontSize: false,
},
})
.useTailwindColors()
.useTailwindFontFamily()
.useTailwindFontSize()
.enable()
};
```
### Relevant .budfiles
_No response_
|
1.0
|
[bug] bud-sass loader not registering, ignores files using a transpiler source - ### Agreement
- [X] This is not a duplicate of an existing issue
- [X] I have read the [guidelines for Contributing to Roots Projects](https://github.com/roots/.github/blob/master/CONTRIBUTING.md)
- [X] This is not a personal support request that should be posted on the [Roots Discourse](https://discourse.roots.io/) community
### Describe the issue
I have a need to load styles/js from other directories than `./resources`. I wish to compile .scss files in a directory to their own CSS files.
I have created a test directory in the root of my sage theme, containing an empty `script.js` and an empty `style.scss`
I then add the test folder to the entry point:
```js
app
/**
* Application entrypoints
*/
.entry({
app: ["@scripts/app", "@styles/app"],
editor: ["@scripts/editor", "@styles/editor"],
test: [app.path("./test/script.js"), app.path("./test/style.scss")]
})
```
After that, I add a transpiler source to the js and sass rules:
```js
app.build.rules.js.setInclude([
bud => bud.path('@src'),
bud => bud.path('./test'),
])
```
This works fine, but the .scss file is ignored. So I go ahead and try add another source to the sass rule:
```js
app.build.rules.sass.setInclude([
bud => bud.path('@src'),
bud => bud.path('./test'),
])
```
Which throws an error:
` Type Error: Cannot read properties of undefined (reading 'setInclude') `
In my bud config, I console log out the `app.build.rules` and see there is no sass rule:
```js
{
yml: Rule {
_app: [Function (anonymous)],
include: [ [Function (anonymous)] ],
test: [Function (anonymous)],
use: [ 'yml' ]
},
webp: Rule {
_app: [Function (anonymous)],
test: [Function (anonymous)],
include: [ [Function (anonymous)] ],
type: 'asset/resource',
generator: [Function (anonymous)]
},
svg: Rule {
_app: [Function (anonymous)],
test: [Function (anonymous)],
include: [ [Function (anonymous)] ],
type: 'asset/resource',
generator: [Function: bound svgGenerator]
},
json: Rule {
_app: [Function (anonymous)],
type: 'json',
include: [ [Function (anonymous)] ],
test: [Function (anonymous)],
parser: [Function (anonymous)]
},
js: Rule {
_app: [Function (anonymous)],
test: [Function (anonymous)],
include: [ [Function (anonymous)] ],
use: [ [Item] ]
},
image: Rule {
_app: [Function (anonymous)],
test: [Function (anonymous)],
include: [ [Function (anonymous)] ],
type: 'asset/resource',
generator: [Function (anonymous)]
},
html: Rule {
_app: [Function (anonymous)],
include: [ [Function (anonymous)] ],
test: [Function (anonymous)],
use: [ 'html' ]
},
font: Rule {
_app: [Function (anonymous)],
type: 'asset',
test: [Function (anonymous)],
include: [ [Function (anonymous)] ],
generator: [Function (anonymous)]
},
cssModule: Rule {
_app: [Function (anonymous)],
test: [Function (anonymous)],
include: [ [Function (anonymous)] ],
use: [ 'precss', 'cssModule', 'postcss' ]
},
css: Rule {
_app: [Function (anonymous)],
test: [Function (anonymous)],
include: [ [Function (anonymous)] ],
use: [ 'precss', 'css', 'postcss' ]
}
}
```
### Expected Behavior
.scss files in additional entry points, that have a transpiler source - to be parsed using bud-sass
### Actual Behavior
This error is thrown: `Type Error: Cannot read properties of undefined (reading 'setInclude')`
### Steps To Reproduce
1. Create a test folder in the root of the teme
2. Add a blank script.js and style.scss to the folder
3. Add the entry points for the test folder into the bud config
4. Add transpiler sources for the js/sass rules
### version
6.4.4
### What package manager are you using?
npm
### version
8.15.0
### Logs
```zsh
Type Error: Cannot read properties of undefined (reading 'setInclude')
at default (file:///Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/bud.config.mjs:19:24)
at Configuration.run (file:///Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/node_modules/@roots/bud-framework/lib/configuration/configuration.js:30:26)
at file:///Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/node_modules/@roots/bud-framework/lib/configuration/index.js:14:29
at Array.map (<anonymous>)
at Module.process (file:///Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/node_modules/@roots/bud-framework/lib/configuration/index.js:13:60)
at Bud.run (file:///Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/node_modules/@roots/bud-framework/lib/methods/run.js:5:29)
at BuildDevelopmentCommand.runCommand (file:///Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/node_modules/@roots/bud/lib/cli/commands/build.base.js:192:24)
at BuildDevelopmentCommand.execute (file:///Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/node_modules/@roots/bud/lib/cli/commands/base.js:179:24)
at async BuildDevelopmentCommand.validateAndExecute (/Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/node_modules/clipanion/lib/advanced/Command.js:73:26)
at async Cli.run (/Users/marcbroad/Desktop/www/sallys/site/web/app/themes/sage/node_modules/clipanion/lib/advanced/Cli.js:222:24)
```
### Configuration
```zsh
// @ts-check
/**
* Build configuration
*
* @see {@link https://bud.js.org/guides/configure}
* @param {import('@roots/bud').Bud} app
*/
export default async (app) => {
console.log(app.build.rules)
app.build.rules.js.setInclude([
bud => bud.path('@src'),
bud => bud.path('./test'),
])
app.build.rules.scss.setInclude([
bud => bud.path('@src'),
bud => bud.path('./test'),
])
app
/**
* Application entrypoints
*/
.entry({
app: ["@scripts/app", "@styles/app"],
editor: ["@scripts/editor", "@styles/editor"],
test: [app.path("./test/script.js"), app.path("./test/style.scss")]
})
/**
* Directory contents to be included in the compilation
*/
.assets(["images"])
/**
* Matched files trigger a page reload when modified
*/
.watch(["resources/views/**/*", "app/**/*"])
/**
* Proxy origin (`WP_HOME`)
*/
.proxy("http://example.test")
/**
* Development origin
*/
.serve("http://0.0.0.0:3000")
/**
* URI of the `public` directory
*/
.setPublicPath("/app/themes/sage/public/")
/**
* Generate WordPress `theme.json`
*
* @note This overwrites `theme.json` on every build.
*/
.wpjson
.settings({
color: {
custom: false,
customGradient: false,
defaultPalette: false,
defaultGradients: false,
},
custom: {
spacing: {},
typography: {
'font-size': {},
'line-height': {},
},
},
spacing: {
padding: true,
units: ['px', '%', 'em', 'rem', 'vw', 'vh'],
},
typography: {
customFontSize: false,
},
})
.useTailwindColors()
.useTailwindFontFamily()
.useTailwindFontSize()
.enable()
};
```
### Relevant .budfiles
_No response_
|
non_process
|
bud sass loader not registering ignores files using a transpiler source agreement this is not a duplicate of an existing issue i have read the this is not a personal support request that should be posted on the community describe the issue i have a need to load styles js from other directories than resources i wish to compile scss files in a directory to their own css files i have created a test directory in the root of my sage theme containing an empty script js and an empty style scss i then add the test folder to the entry point js app application entrypoints entry app editor test after that i add a transpiler source to the js and sass rules js app build rules js setinclude bud bud path src bud bud path test this works fine but the scss file is ignored so i go ahead and try add another source to the sass rule js app build rules sass setinclude bud bud path src bud bud path test which throws an error type error cannot read properties of undefined reading setinclude in my bud config i console log out the app build rules and see there is no sass rule js yml rule app include test use webp rule app test include type asset resource generator svg rule app test include type asset resource generator json rule app type json include test parser js rule app test include use image rule app test include type asset resource generator html rule app include test use font rule app type asset test include generator cssmodule rule app test include use css rule app test include use expected behavior scss files in additional entry points that have a transpiler source to be parsed using bud sass actual behavior this error is thrown type error cannot read properties of undefined reading setinclude steps to reproduce create a test folder in the root of the teme add a blank script js and style scss to the folder add the entry points for the test folder into the bud config add transpiler sources for the js sass rules version what package manager are you using npm version logs zsh type error cannot read properties of undefined reading setinclude at default file users marcbroad desktop www sallys site web app themes sage bud config mjs at configuration run file users marcbroad desktop www sallys site web app themes sage node modules roots bud framework lib configuration configuration js at file users marcbroad desktop www sallys site web app themes sage node modules roots bud framework lib configuration index js at array map at module process file users marcbroad desktop www sallys site web app themes sage node modules roots bud framework lib configuration index js at bud run file users marcbroad desktop www sallys site web app themes sage node modules roots bud framework lib methods run js at builddevelopmentcommand runcommand file users marcbroad desktop www sallys site web app themes sage node modules roots bud lib cli commands build base js at builddevelopmentcommand execute file users marcbroad desktop www sallys site web app themes sage node modules roots bud lib cli commands base js at async builddevelopmentcommand validateandexecute users marcbroad desktop www sallys site web app themes sage node modules clipanion lib advanced command js at async cli run users marcbroad desktop www sallys site web app themes sage node modules clipanion lib advanced cli js configuration zsh ts check build configuration see link param import roots bud bud app export default async app console log app build rules app build rules js setinclude bud bud path src bud bud path test app build rules scss setinclude bud bud path src bud bud path test app application entrypoints entry app editor test directory contents to be included in the compilation assets matched files trigger a page reload when modified watch proxy origin wp home proxy development origin serve uri of the public directory setpublicpath app themes sage public generate wordpress theme json note this overwrites theme json on every build wpjson settings color custom false customgradient false defaultpalette false defaultgradients false custom spacing typography font size line height spacing padding true units typography customfontsize false usetailwindcolors usetailwindfontfamily usetailwindfontsize enable relevant budfiles no response
| 0
|
11,419
| 14,246,202,566
|
IssuesEvent
|
2020-11-19 09:45:10
|
freedomofpress/securedrop
|
https://api.github.com/repos/freedomofpress/securedrop
|
closed
|
i18n_tool.py list-translators is broken by multiple syncs in a release
|
goals: speed up release process
|
## Description
The `list-translators` function has a brittle heuristic for determining when to start gathering translator contributions; it looks backward through the git history for the most recent `l10n: sync` message. If we have to sync more than once for a release, to incorporate source string feedback or whatever, then it will miss contributions since the last release and before that sync.
|
1.0
|
i18n_tool.py list-translators is broken by multiple syncs in a release - ## Description
The `list-translators` function has a brittle heuristic for determining when to start gathering translator contributions; it looks backward through the git history for the most recent `l10n: sync` message. If we have to sync more than once for a release, to incorporate source string feedback or whatever, then it will miss contributions since the last release and before that sync.
|
process
|
tool py list translators is broken by multiple syncs in a release description the list translators function has a brittle heuristic for determining when to start gathering translator contributions it looks backward through the git history for the most recent sync message if we have to sync more than once for a release to incorporate source string feedback or whatever then it will miss contributions since the last release and before that sync
| 1
|
8,292
| 11,458,273,057
|
IssuesEvent
|
2020-02-07 02:44:53
|
googleapis/google-cloud-python
|
https://api.github.com/repos/googleapis/google-cloud-python
|
closed
|
Generated noxfiles missing 'cover' session
|
api: automl api: bigquerydatatransfer api: cloudasset api: cloudiot api: cloudkms api: cloudtasks api: container api: dataproc api: oslogin api: texttospeech api: videointelligence testing type: process
|
They are also missing the `lint` session.
While the argument might be made, "they are autogen-only", the fact is that a) the `gapic-generator`, `synthtool`, etc. are software, and have bugs too; b) the local `synth.py` is perfectly capable of injecting its *own* bugs.
`videointelligence` is notable in this list because it is at "beta" support level, so we should expect higher code quality for it.
|
1.0
|
Generated noxfiles missing 'cover' session - They are also missing the `lint` session.
While the argument might be made, "they are autogen-only", the fact is that a) the `gapic-generator`, `synthtool`, etc. are software, and have bugs too; b) the local `synth.py` is perfectly capable of injecting its *own* bugs.
`videointelligence` is notable in this list because it is at "beta" support level, so we should expect higher code quality for it.
|
process
|
generated noxfiles missing cover session they are also missing the lint session while the argument might be made they are autogen only the fact is that a the gapic generator synthtool etc are software and have bugs too b the local synth py is perfectly capable of injecting its own bugs videointelligence is notable in this list because it is at beta support level so we should expect higher code quality for it
| 1
|
659,223
| 21,919,625,872
|
IssuesEvent
|
2022-05-22 11:21:59
|
kubernetes/ingress-nginx
|
https://api.github.com/repos/kubernetes/ingress-nginx
|
closed
|
Allow ingress controller to set default annotations for ingress resources
|
kind/feature lifecycle/rotten needs-triage needs-priority
|
When configuring an ingress controller we can set some configuration via the config map. In Helm we can use the `controlller.config` to pass the configuration to avoid having to set those setting as annotations in the ingress resource.
It would be nice to also have a way to set some default annotations that all ingress resources that are using that particular `ingressClass` will inherit.
For example `cert-manager.io/cluster-issuer: somevalue` could be a useful annotation to add to the ingress controller so that ingress resources don't require the annotation to be set.
Is this behaviour already available somehow? Please let me know if I need to clarify the issue a bit more. THanks.
|
1.0
|
Allow ingress controller to set default annotations for ingress resources - When configuring an ingress controller we can set some configuration via the config map. In Helm we can use the `controlller.config` to pass the configuration to avoid having to set those setting as annotations in the ingress resource.
It would be nice to also have a way to set some default annotations that all ingress resources that are using that particular `ingressClass` will inherit.
For example `cert-manager.io/cluster-issuer: somevalue` could be a useful annotation to add to the ingress controller so that ingress resources don't require the annotation to be set.
Is this behaviour already available somehow? Please let me know if I need to clarify the issue a bit more. THanks.
|
non_process
|
allow ingress controller to set default annotations for ingress resources when configuring an ingress controller we can set some configuration via the config map in helm we can use the controlller config to pass the configuration to avoid having to set those setting as annotations in the ingress resource it would be nice to also have a way to set some default annotations that all ingress resources that are using that particular ingressclass will inherit for example cert manager io cluster issuer somevalue could be a useful annotation to add to the ingress controller so that ingress resources don t require the annotation to be set is this behaviour already available somehow please let me know if i need to clarify the issue a bit more thanks
| 0
|
18,926
| 24,880,715,634
|
IssuesEvent
|
2022-10-28 00:37:40
|
apache/arrow-rs
|
https://api.github.com/repos/apache/arrow-rs
|
closed
|
Make take kernel not take values of childs when taking a null
|
arrow development-process
|
*Note*: migrated from original JIRA: https://issues.apache.org/jira/browse/ARROW-10594
Currently, take just takes all values from the childs, irrespectively of whether we took a null or not.
|
1.0
|
Make take kernel not take values of childs when taking a null - *Note*: migrated from original JIRA: https://issues.apache.org/jira/browse/ARROW-10594
Currently, take just takes all values from the childs, irrespectively of whether we took a null or not.
|
process
|
make take kernel not take values of childs when taking a null note migrated from original jira currently take just takes all values from the childs irrespectively of whether we took a null or not
| 1
|
418,056
| 28,113,374,333
|
IssuesEvent
|
2023-03-31 08:56:10
|
tangphi/ped
|
https://api.github.com/repos/tangphi/ped
|
opened
|
Not enough visuals in UG
|
severity.VeryLow type.DocumentationBug
|

Since the graph feature outputs a graph, it would be nice to have an example output based on the given example input.
<!--session: 1680252431019-08cca115-f668-44e8-9c9b-d3d5514bd741-->
<!--Version: Web v3.4.7-->
|
1.0
|
Not enough visuals in UG - 
Since the graph feature outputs a graph, it would be nice to have an example output based on the given example input.
<!--session: 1680252431019-08cca115-f668-44e8-9c9b-d3d5514bd741-->
<!--Version: Web v3.4.7-->
|
non_process
|
not enough visuals in ug since the graph feature outputs a graph it would be nice to have an example output based on the given example input
| 0
|
22,391
| 31,142,286,668
|
IssuesEvent
|
2023-08-16 01:44:19
|
cypress-io/cypress
|
https://api.github.com/repos/cypress-io/cypress
|
closed
|
Flaky test: `cy.task('__internal_scaffoldProject')` failed with the following error: > EPERM: operation not permitted, stat 'C:\Users\circleci\AppData\Local\Temp\cy-projects\cypress-in-cypress'
|
OS: windows stage: backlog process: flaky test topic: flake ❄️ topic: scaffoldProject stale
|
### Link to dashboard or CircleCI failure
https://app.circleci.com/pipelines/github/cypress-io/cypress/41757/workflows/1a0d6f2e-ac67-4ac6-ab24-2c00e7149ea4/jobs/1730826/tests#failed-test-0
### Link to failing test in GitHub
N/A
### Analysis
<img width="1102" alt="Screen Shot 2022-08-11 at 7 55 03 PM" src="https://user-images.githubusercontent.com/26726429/184276453-d4589029-aed7-48ff-8953-7d796a8bd762.png">
### Cypress Version
10.4.0
### Other
_No response_
|
1.0
|
Flaky test: `cy.task('__internal_scaffoldProject')` failed with the following error: > EPERM: operation not permitted, stat 'C:\Users\circleci\AppData\Local\Temp\cy-projects\cypress-in-cypress' - ### Link to dashboard or CircleCI failure
https://app.circleci.com/pipelines/github/cypress-io/cypress/41757/workflows/1a0d6f2e-ac67-4ac6-ab24-2c00e7149ea4/jobs/1730826/tests#failed-test-0
### Link to failing test in GitHub
N/A
### Analysis
<img width="1102" alt="Screen Shot 2022-08-11 at 7 55 03 PM" src="https://user-images.githubusercontent.com/26726429/184276453-d4589029-aed7-48ff-8953-7d796a8bd762.png">
### Cypress Version
10.4.0
### Other
_No response_
|
process
|
flaky test cy task internal scaffoldproject failed with the following error eperm operation not permitted stat c users circleci appdata local temp cy projects cypress in cypress link to dashboard or circleci failure link to failing test in github n a analysis img width alt screen shot at pm src cypress version other no response
| 1
|
70,187
| 13,436,035,455
|
IssuesEvent
|
2020-09-07 13:48:21
|
easably/website
|
https://api.github.com/repos/easably/website
|
opened
|
Design two pop-up window for standard Promo Code.
|
design promo codes
|
Description:
Develop two pop-up windows for standard Promo Code.
The first layout is a pop-up window after choosing a free promo code (Apply Now).
The second layout is the page that the customer will see after confirming the application of the promo code (Congratulation+ Your Promo code has been successfully applied).
An example for design can be taken from the page: https://www.figma.com/file/oRhHhpPn92aPwa8kSFxsJ6/Untitled?node-id=0%3A1
|
1.0
|
Design two pop-up window for standard Promo Code. - Description:
Develop two pop-up windows for standard Promo Code.
The first layout is a pop-up window after choosing a free promo code (Apply Now).
The second layout is the page that the customer will see after confirming the application of the promo code (Congratulation+ Your Promo code has been successfully applied).
An example for design can be taken from the page: https://www.figma.com/file/oRhHhpPn92aPwa8kSFxsJ6/Untitled?node-id=0%3A1
|
non_process
|
design two pop up window for standard promo code description develop two pop up windows for standard promo code the first layout is a pop up window after choosing a free promo code apply now the second layout is the page that the customer will see after confirming the application of the promo code congratulation your promo code has been successfully applied an example for design can be taken from the page
| 0
|
181,961
| 14,894,871,680
|
IssuesEvent
|
2021-01-21 08:16:28
|
equinor/flownet
|
https://api.github.com/repos/equinor/flownet
|
opened
|
Write documentation on relperm modelling
|
documentation
|
The relative permeability modelling has in the meanwhile become quite extensive. We should write documentation on it.
|
1.0
|
Write documentation on relperm modelling - The relative permeability modelling has in the meanwhile become quite extensive. We should write documentation on it.
|
non_process
|
write documentation on relperm modelling the relative permeability modelling has in the meanwhile become quite extensive we should write documentation on it
| 0
|
18,531
| 24,552,697,614
|
IssuesEvent
|
2022-10-12 13:44:55
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[Mobile apps] Study activities are not getting loaded for enrolled studies
|
Bug Blocker P0 iOS Android Process: Fixed Process: Tested dev
|
Study activities are not getting loaded for enrolled studies - Getting continuous loading

|
2.0
|
[Mobile apps] Study activities are not getting loaded for enrolled studies - Study activities are not getting loaded for enrolled studies - Getting continuous loading

|
process
|
study activities are not getting loaded for enrolled studies study activities are not getting loaded for enrolled studies getting continuous loading
| 1
|
21,741
| 30,257,820,351
|
IssuesEvent
|
2023-07-07 05:20:34
|
open-telemetry/opentelemetry-collector-contrib
|
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
|
closed
|
Grab docker label using docker/resourcedetection
|
enhancement Stale processor/resourcedetection closed as inactive
|
### Component(s)
processor/resourcedetection
### Is your feature request related to a problem? Please describe.
I'm using nomad to run otel, I got difficulty when trying to get the nomad label and send it to loki.
The possible way to grab nomad label is by using docker/resourcedetection. but currently, docker/resourcedetection only supports [host.name and os.type](https://github.com/open-telemetry/opentelemetry-collector-contrib/blob/main/processor/resourcedetectionprocessor/README.md#system-metadata)
Here are some label from containers that might we can use to create our own label
```
"Labels": {
"com.hashicorp.nomad.alloc_id": "40c4c3b8-1d1f-8f58-94ca-4efda336d62e",
"com.hashicorp.nomad.job_id": "fabio",
"com.hashicorp.nomad.job_name": "fabio",
"com.hashicorp.nomad.namespace": "default",
"com.hashicorp.nomad.node_id": "1610be6d-eea9-c4da-8d88-b945fe6e3306",
"com.hashicorp.nomad.node_name": "t3a-small-ip-10-11-82-122-aws-client",
"com.hashicorp.nomad.task_group_name": "fabio",
"com.hashicorp.nomad.task_name": "fabio",
"org.opencontainers.image.created": "2022-09-10T19:11:50Z",
"org.opencontainers.image.revision": "ebf15c22df1a4b4b367a9043865d3fe51d5db013",
"org.opencontainers.image.title": "fabio",
"org.opencontainers.image.version": "1.6.2"
}
```
### Describe the solution you'd like
Add support to use docker label in docker/resourcedetection like [vector (see vector configuration file)](https://atodorov.me/2021/07/09/logging-on-nomad-and-log-aggregation-with-loki/)
```
resourcedetection/docker:
detectors: [env, docker]
timeout: 2s
override: false
attributes:
- host.name
- os.type
- label.com.hashicorp.nomad.job_name
- label.com.hashicorp.nomad.task_name
```
### Describe alternatives you've considered
-
### Additional context
-
|
1.0
|
Grab docker label using docker/resourcedetection - ### Component(s)
processor/resourcedetection
### Is your feature request related to a problem? Please describe.
I'm using nomad to run otel, I got difficulty when trying to get the nomad label and send it to loki.
The possible way to grab nomad label is by using docker/resourcedetection. but currently, docker/resourcedetection only supports [host.name and os.type](https://github.com/open-telemetry/opentelemetry-collector-contrib/blob/main/processor/resourcedetectionprocessor/README.md#system-metadata)
Here are some label from containers that might we can use to create our own label
```
"Labels": {
"com.hashicorp.nomad.alloc_id": "40c4c3b8-1d1f-8f58-94ca-4efda336d62e",
"com.hashicorp.nomad.job_id": "fabio",
"com.hashicorp.nomad.job_name": "fabio",
"com.hashicorp.nomad.namespace": "default",
"com.hashicorp.nomad.node_id": "1610be6d-eea9-c4da-8d88-b945fe6e3306",
"com.hashicorp.nomad.node_name": "t3a-small-ip-10-11-82-122-aws-client",
"com.hashicorp.nomad.task_group_name": "fabio",
"com.hashicorp.nomad.task_name": "fabio",
"org.opencontainers.image.created": "2022-09-10T19:11:50Z",
"org.opencontainers.image.revision": "ebf15c22df1a4b4b367a9043865d3fe51d5db013",
"org.opencontainers.image.title": "fabio",
"org.opencontainers.image.version": "1.6.2"
}
```
### Describe the solution you'd like
Add support to use docker label in docker/resourcedetection like [vector (see vector configuration file)](https://atodorov.me/2021/07/09/logging-on-nomad-and-log-aggregation-with-loki/)
```
resourcedetection/docker:
detectors: [env, docker]
timeout: 2s
override: false
attributes:
- host.name
- os.type
- label.com.hashicorp.nomad.job_name
- label.com.hashicorp.nomad.task_name
```
### Describe alternatives you've considered
-
### Additional context
-
|
process
|
grab docker label using docker resourcedetection component s processor resourcedetection is your feature request related to a problem please describe i m using nomad to run otel i got difficulty when trying to get the nomad label and send it to loki the possible way to grab nomad label is by using docker resourcedetection but currently docker resourcedetection only supports here are some label from containers that might we can use to create our own label labels com hashicorp nomad alloc id com hashicorp nomad job id fabio com hashicorp nomad job name fabio com hashicorp nomad namespace default com hashicorp nomad node id com hashicorp nomad node name small ip aws client com hashicorp nomad task group name fabio com hashicorp nomad task name fabio org opencontainers image created org opencontainers image revision org opencontainers image title fabio org opencontainers image version describe the solution you d like add support to use docker label in docker resourcedetection like resourcedetection docker detectors timeout override false attributes host name os type label com hashicorp nomad job name label com hashicorp nomad task name describe alternatives you ve considered additional context
| 1
|
233,137
| 25,738,940,407
|
IssuesEvent
|
2022-12-08 03:59:30
|
CDCgov/prime-reportstream
|
https://api.github.com/repos/CDCgov/prime-reportstream
|
closed
|
React App idle timer logout
|
security experience
|
The react app is a full SPA and pages load without making https requests to the server.
While the REST API can detect "idle" time by delays between requests and/or can be controlled by Okta configuration(?), the user experience isn't the greatest. At some point the requests to get data would just start failing.
A better approach is to use a idle detect hook that warns before idle timeout. Then after the logged out, a notice can be displayed explaining what happened.
This might be a security requirement? We should verify.
|
True
|
React App idle timer logout - The react app is a full SPA and pages load without making https requests to the server.
While the REST API can detect "idle" time by delays between requests and/or can be controlled by Okta configuration(?), the user experience isn't the greatest. At some point the requests to get data would just start failing.
A better approach is to use a idle detect hook that warns before idle timeout. Then after the logged out, a notice can be displayed explaining what happened.
This might be a security requirement? We should verify.
|
non_process
|
react app idle timer logout the react app is a full spa and pages load without making https requests to the server while the rest api can detect idle time by delays between requests and or can be controlled by okta configuration the user experience isn t the greatest at some point the requests to get data would just start failing a better approach is to use a idle detect hook that warns before idle timeout then after the logged out a notice can be displayed explaining what happened this might be a security requirement we should verify
| 0
|
7,993
| 2,611,071,447
|
IssuesEvent
|
2015-02-27 00:33:21
|
alistairreilly/andors-trail
|
https://api.github.com/repos/alistairreilly/andors-trail
|
opened
|
combat doesn't end after monster dies
|
auto-migrated Type-Defect
|
```
Before posting, please read the following guidelines for posts in the issue
tracker:
http://code.google.com/p/andors-trail/wiki/Forums_vs_issuetracker
What steps will reproduce the problem?
1.I have gutherbeards dagger equipped
2.enemy loses last hp points from bleeding curse
3.I have to complete another round of attack against a dead enemy before the
fight is over
What is the expected output? What do you see instead? I expect the fight to end
once the enemy is dead
What version of the product are you using? On what device? V 0.6.12 on samsung
galaxy s2 sprint
Please provide any additional information below.
```
Original issue reported on code.google.com by `blue...@gmail.com` on 26 Jun 2013 at 4:38
|
1.0
|
combat doesn't end after monster dies - ```
Before posting, please read the following guidelines for posts in the issue
tracker:
http://code.google.com/p/andors-trail/wiki/Forums_vs_issuetracker
What steps will reproduce the problem?
1.I have gutherbeards dagger equipped
2.enemy loses last hp points from bleeding curse
3.I have to complete another round of attack against a dead enemy before the
fight is over
What is the expected output? What do you see instead? I expect the fight to end
once the enemy is dead
What version of the product are you using? On what device? V 0.6.12 on samsung
galaxy s2 sprint
Please provide any additional information below.
```
Original issue reported on code.google.com by `blue...@gmail.com` on 26 Jun 2013 at 4:38
|
non_process
|
combat doesn t end after monster dies before posting please read the following guidelines for posts in the issue tracker what steps will reproduce the problem i have gutherbeards dagger equipped enemy loses last hp points from bleeding curse i have to complete another round of attack against a dead enemy before the fight is over what is the expected output what do you see instead i expect the fight to end once the enemy is dead what version of the product are you using on what device v on samsung galaxy sprint please provide any additional information below original issue reported on code google com by blue gmail com on jun at
| 0
|
18,007
| 24,024,214,291
|
IssuesEvent
|
2022-09-15 10:06:01
|
anitsh/til
|
https://api.github.com/repos/anitsh/til
|
opened
|
Guiding principle: cross-pollination over imposed standards
|
principle practice blog protocol process
|
Standards are useful to simplify learning and address variation of performance.
Standards are useful to avoid everyone having to learn a new way of doing things every time they interact with a new team.
Non-standard team interaction protocols

Standards are useful to address variation of performance, that is, if there’s a better way of doing something, a standard can be used to spread it across teams.
Re-inventing the wheel rather than spreading it via a standard
The problem with imposed standards is context
There are problems with imposing a centralised standard.
There’s a cost to having to learn different approaches every time you switch context BUT it’s also unlikely that one approach is optimal for every context AND we want to allow for experimentation to discover even better approaches.
Make defaults easy but leave the option open for alternatives
Cross-pollination encourages de facto standardisation while allowing for flexibility
Ensure everyone is aware of defaults but also has the autonomy to choose an alternate approach as appropriate.
Make defaults easy to do (aka [Golden Path](https://engineering.atspotify.com/2020/08/how-we-use-golden-paths-to-solve-fragmentation-in-our-software-ecosystem/)) but leave the option open to choose an alternate approach as appropriate. This allows for context-specific adjustments AND experimentation to find even better approaches.

Cross-boundary protocols generally warrants stronger guidance, even imposition
Communication and interaction protocols across boundaries are where standardisation is important even if it requires more imposition.

# Resource
- https://jchyip.medium.com/guiding-principle-cross-pollination-over-imposed-standards-a2375d0e8de6
|
1.0
|
Guiding principle: cross-pollination over imposed standards - Standards are useful to simplify learning and address variation of performance.
Standards are useful to avoid everyone having to learn a new way of doing things every time they interact with a new team.
Non-standard team interaction protocols

Standards are useful to address variation of performance, that is, if there’s a better way of doing something, a standard can be used to spread it across teams.
Re-inventing the wheel rather than spreading it via a standard
The problem with imposed standards is context
There are problems with imposing a centralised standard.
There’s a cost to having to learn different approaches every time you switch context BUT it’s also unlikely that one approach is optimal for every context AND we want to allow for experimentation to discover even better approaches.
Make defaults easy but leave the option open for alternatives
Cross-pollination encourages de facto standardisation while allowing for flexibility
Ensure everyone is aware of defaults but also has the autonomy to choose an alternate approach as appropriate.
Make defaults easy to do (aka [Golden Path](https://engineering.atspotify.com/2020/08/how-we-use-golden-paths-to-solve-fragmentation-in-our-software-ecosystem/)) but leave the option open to choose an alternate approach as appropriate. This allows for context-specific adjustments AND experimentation to find even better approaches.

Cross-boundary protocols generally warrants stronger guidance, even imposition
Communication and interaction protocols across boundaries are where standardisation is important even if it requires more imposition.

# Resource
- https://jchyip.medium.com/guiding-principle-cross-pollination-over-imposed-standards-a2375d0e8de6
|
process
|
guiding principle cross pollination over imposed standards standards are useful to simplify learning and address variation of performance standards are useful to avoid everyone having to learn a new way of doing things every time they interact with a new team non standard team interaction protocols standards are useful to address variation of performance that is if there’s a better way of doing something a standard can be used to spread it across teams re inventing the wheel rather than spreading it via a standard the problem with imposed standards is context there are problems with imposing a centralised standard there’s a cost to having to learn different approaches every time you switch context but it’s also unlikely that one approach is optimal for every context and we want to allow for experimentation to discover even better approaches make defaults easy but leave the option open for alternatives cross pollination encourages de facto standardisation while allowing for flexibility ensure everyone is aware of defaults but also has the autonomy to choose an alternate approach as appropriate make defaults easy to do aka but leave the option open to choose an alternate approach as appropriate this allows for context specific adjustments and experimentation to find even better approaches cross boundary protocols generally warrants stronger guidance even imposition communication and interaction protocols across boundaries are where standardisation is important even if it requires more imposition resource
| 1
|
52,193
| 7,752,279,930
|
IssuesEvent
|
2018-05-30 19:46:41
|
amberframework/granite
|
https://api.github.com/repos/amberframework/granite
|
closed
|
Validation helpers
|
kind:documentation kind:enhancement pr:needs-review
|
As of now the validators usage is not documented in the readme. We should do that.
|
1.0
|
Validation helpers - As of now the validators usage is not documented in the readme. We should do that.
|
non_process
|
validation helpers as of now the validators usage is not documented in the readme we should do that
| 0
|
16,663
| 21,731,104,807
|
IssuesEvent
|
2022-05-11 12:06:09
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Crash when running check validity using QGIS method
|
Processing Bug Crash/Data Corruption
|
### What is the bug or the crash?
When running "Check Validity" on a particular geopackage using the QGIS method, QGIS crashes. When using the GEOS method, it reports no errors. I've also tested on a clean user profile, and the crash still occurs.
I've isolated the problem to this single multi-polygon feature in a geopackage:
[checkvalidity_crash.zip](https://github.com/qgis/QGIS/files/8663128/checkvalidity_crash.zip)
## Report Details
**Python Stack Trace**
```
Windows fatal exception: access violation
Current thread 0x00024620 (most recent call first):
File "C:\OSGeo4W/apps/qgis/./python/plugins\processing\algs\qgis\CheckValidity.py", line 160 in doCheck
errors = list(geom.validateGeometry(Qgis.GeometryValidationEngine(method), flags))
File "C:\OSGeo4W/apps/qgis/./python/plugins\processing\algs\qgis\CheckValidity.py", line 124 in processAlgorithm
return self.doCheck(
Thread 0x00023220 (most recent call first):
File "C:\OSGeo4W/apps/qgis/./python/plugins\processing\ProcessingPlugin.py", line 395 in executeAlgorithm
dlg.exec_()
File "C:\OSGeo4W/apps/qgis/./python/plugins\processing\gui\ProcessingToolbox.py", line 234 in executeAlgorithm
self.executeWithGui.emit(alg.id(), self, self.in_place_mode, False)
File "C:\OSGeo4W/apps/qgis/./python/plugins\processing\gui\AlgorithmLocatorFilter.py", line 120 in triggerResult
dlg.exec_()
```
**Stack Trace**
No stack trace is available.
**QGIS Info**
QGIS Version: 3.24.2-Tisler
QGIS code revision: 13c1a028
Compiled against Qt: 5.15.2
Running against Qt: 5.15.2
Compiled against GDAL: 3.4.2
Running against GDAL: 3.4.2
**System Info**
CPU Type: x86_64
Kernel Type: winnt
Kernel Version: 10.0.19044
### Steps to reproduce the issue
1. Open this geopackage [checkvalidity_crash.zip](https://github.com/qgis/QGIS/files/8663128/checkvalidity_crash.zip) in QGIS
2. Run the processing tool "Check Validity" with method = QGIS
3. Crash!
### Versions
QGIS version
3.24.2-Tisler
QGIS code revision
13c1a028
Qt version
5.15.2
Python version
3.9.5
GDAL/OGR version
3.4.2
PROJ version
9.0.0
EPSG Registry database version
v10.054 (2022-02-13)
GEOS version
3.10.2-CAPI-1.16.0
SQLite version
3.38.1
PDAL version
2.3.0
PostgreSQL client version
unknown
SpatiaLite version
5.0.1
QWT version
6.1.3
QScintilla2 version
2.11.5
OS version
Windows 10 Version 2009
Active Python plugins
db_manager
0.1.20
grassprovider
2.12.99
MetaSearch
0.3.6
processing
2.12.99
sagaprovider
2.12.99
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [X] I tried with a new QGIS profile
### Additional context
_No response_
|
1.0
|
Crash when running check validity using QGIS method - ### What is the bug or the crash?
When running "Check Validity" on a particular geopackage using the QGIS method, QGIS crashes. When using the GEOS method, it reports no errors. I've also tested on a clean user profile, and the crash still occurs.
I've isolated the problem to this single multi-polygon feature in a geopackage:
[checkvalidity_crash.zip](https://github.com/qgis/QGIS/files/8663128/checkvalidity_crash.zip)
## Report Details
**Python Stack Trace**
```
Windows fatal exception: access violation
Current thread 0x00024620 (most recent call first):
File "C:\OSGeo4W/apps/qgis/./python/plugins\processing\algs\qgis\CheckValidity.py", line 160 in doCheck
errors = list(geom.validateGeometry(Qgis.GeometryValidationEngine(method), flags))
File "C:\OSGeo4W/apps/qgis/./python/plugins\processing\algs\qgis\CheckValidity.py", line 124 in processAlgorithm
return self.doCheck(
Thread 0x00023220 (most recent call first):
File "C:\OSGeo4W/apps/qgis/./python/plugins\processing\ProcessingPlugin.py", line 395 in executeAlgorithm
dlg.exec_()
File "C:\OSGeo4W/apps/qgis/./python/plugins\processing\gui\ProcessingToolbox.py", line 234 in executeAlgorithm
self.executeWithGui.emit(alg.id(), self, self.in_place_mode, False)
File "C:\OSGeo4W/apps/qgis/./python/plugins\processing\gui\AlgorithmLocatorFilter.py", line 120 in triggerResult
dlg.exec_()
```
**Stack Trace**
No stack trace is available.
**QGIS Info**
QGIS Version: 3.24.2-Tisler
QGIS code revision: 13c1a028
Compiled against Qt: 5.15.2
Running against Qt: 5.15.2
Compiled against GDAL: 3.4.2
Running against GDAL: 3.4.2
**System Info**
CPU Type: x86_64
Kernel Type: winnt
Kernel Version: 10.0.19044
### Steps to reproduce the issue
1. Open this geopackage [checkvalidity_crash.zip](https://github.com/qgis/QGIS/files/8663128/checkvalidity_crash.zip) in QGIS
2. Run the processing tool "Check Validity" with method = QGIS
3. Crash!
### Versions
QGIS version
3.24.2-Tisler
QGIS code revision
13c1a028
Qt version
5.15.2
Python version
3.9.5
GDAL/OGR version
3.4.2
PROJ version
9.0.0
EPSG Registry database version
v10.054 (2022-02-13)
GEOS version
3.10.2-CAPI-1.16.0
SQLite version
3.38.1
PDAL version
2.3.0
PostgreSQL client version
unknown
SpatiaLite version
5.0.1
QWT version
6.1.3
QScintilla2 version
2.11.5
OS version
Windows 10 Version 2009
Active Python plugins
db_manager
0.1.20
grassprovider
2.12.99
MetaSearch
0.3.6
processing
2.12.99
sagaprovider
2.12.99
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [X] I tried with a new QGIS profile
### Additional context
_No response_
|
process
|
crash when running check validity using qgis method what is the bug or the crash when running check validity on a particular geopackage using the qgis method qgis crashes when using the geos method it reports no errors i ve also tested on a clean user profile and the crash still occurs i ve isolated the problem to this single multi polygon feature in a geopackage report details python stack trace windows fatal exception access violation current thread most recent call first file c apps qgis python plugins processing algs qgis checkvalidity py line in docheck errors list geom validategeometry qgis geometryvalidationengine method flags file c apps qgis python plugins processing algs qgis checkvalidity py line in processalgorithm return self docheck thread most recent call first file c apps qgis python plugins processing processingplugin py line in executealgorithm dlg exec file c apps qgis python plugins processing gui processingtoolbox py line in executealgorithm self executewithgui emit alg id self self in place mode false file c apps qgis python plugins processing gui algorithmlocatorfilter py line in triggerresult dlg exec stack trace no stack trace is available qgis info qgis version tisler qgis code revision compiled against qt running against qt compiled against gdal running against gdal system info cpu type kernel type winnt kernel version steps to reproduce the issue open this geopackage in qgis run the processing tool check validity with method qgis crash versions qgis version tisler qgis code revision qt version python version gdal ogr version proj version epsg registry database version geos version capi sqlite version pdal version postgresql client version unknown spatialite version qwt version version os version windows version active python plugins db manager grassprovider metasearch processing sagaprovider supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context no response
| 1
|
19,706
| 26,053,298,060
|
IssuesEvent
|
2022-12-22 21:16:39
|
MPMG-DCC-UFMG/C01
|
https://api.github.com/repos/MPMG-DCC-UFMG/C01
|
opened
|
Interface de passos com Vue.js - Tratamento de contextos
|
[1] Bug [0] Desenvolvimento [2] Média Prioridade [3] Processamento Dinâmico
|
## Comportamento Esperado
Um passo dentro do contexto de nova aba ou iframe deve possuir uma cor diferente para indicar esse contexto, conforme era feito na versão anterior da interface de passos. Além disso, deve ser tratado o caso onde uma alteração na interface invalide o contexto de um passo, por exemplo: se há um passo de "fechar aba" após um "abrir em nova aba", e esse passo de "abrir em nova aba" é removido, algo deve ser feito para manter a consistência da configuração (remover o "fechar aba" correspondente, exibir uma mensagem para o usuário indicando da situação, impedir o salvamento do coletor até a correção do problema, etc). Também devem ser tratadas as situações onde o passo que altera o contexto é editado, movido, etc.
## Comportamento Atual
Nenhum tratamento especial é dado aos passos devido ao contexto de execução.
## Sistema
Branch `issue-882`.
|
1.0
|
Interface de passos com Vue.js - Tratamento de contextos - ## Comportamento Esperado
Um passo dentro do contexto de nova aba ou iframe deve possuir uma cor diferente para indicar esse contexto, conforme era feito na versão anterior da interface de passos. Além disso, deve ser tratado o caso onde uma alteração na interface invalide o contexto de um passo, por exemplo: se há um passo de "fechar aba" após um "abrir em nova aba", e esse passo de "abrir em nova aba" é removido, algo deve ser feito para manter a consistência da configuração (remover o "fechar aba" correspondente, exibir uma mensagem para o usuário indicando da situação, impedir o salvamento do coletor até a correção do problema, etc). Também devem ser tratadas as situações onde o passo que altera o contexto é editado, movido, etc.
## Comportamento Atual
Nenhum tratamento especial é dado aos passos devido ao contexto de execução.
## Sistema
Branch `issue-882`.
|
process
|
interface de passos com vue js tratamento de contextos comportamento esperado um passo dentro do contexto de nova aba ou iframe deve possuir uma cor diferente para indicar esse contexto conforme era feito na versão anterior da interface de passos além disso deve ser tratado o caso onde uma alteração na interface invalide o contexto de um passo por exemplo se há um passo de fechar aba após um abrir em nova aba e esse passo de abrir em nova aba é removido algo deve ser feito para manter a consistência da configuração remover o fechar aba correspondente exibir uma mensagem para o usuário indicando da situação impedir o salvamento do coletor até a correção do problema etc também devem ser tratadas as situações onde o passo que altera o contexto é editado movido etc comportamento atual nenhum tratamento especial é dado aos passos devido ao contexto de execução sistema branch issue
| 1
|
49,161
| 6,150,141,237
|
IssuesEvent
|
2017-06-27 21:45:35
|
navx2810/gbs-grm
|
https://api.github.com/repos/navx2810/gbs-grm
|
opened
|
Create flow of application.
|
design
|
The flow is the "flow" of the application from page/component to page/component. Usually done in a diagram.
|
1.0
|
Create flow of application. - The flow is the "flow" of the application from page/component to page/component. Usually done in a diagram.
|
non_process
|
create flow of application the flow is the flow of the application from page component to page component usually done in a diagram
| 0
|
7,903
| 3,633,156,525
|
IssuesEvent
|
2016-02-11 13:28:01
|
catapult-project/catapult
|
https://api.github.com/repos/catapult-project/catapult
|
opened
|
Get rid of <tr-ui-u-time-stamp-span> and <tr-ui-u-time-duration-span>
|
Code Health Good First Bug
|
_[Follow-up for #1981]_
Both elements are unnecessary wrappers around <tt>\<tr-ui-u-scalar-span\></tt>. I propose doing the following:
* Replace <tt>\<tr-ui-u-__time-stamp__-span\></tt> with <tt>\<tr-ui-u-__scalar__-span __unit="timeStampInMs"__\></tt> and
* Replace <tt>tr.ui.units.create**TimeStamp**Span(x)</tt> with <tt>tr.ui.units.create**Scalar**Span(__new tr.b.u.Scalar__(x, __tr.b.u.Unit.byName.timeStampInMs__))</tt>.
@natduca: Sounds good?
|
1.0
|
Get rid of <tr-ui-u-time-stamp-span> and <tr-ui-u-time-duration-span> - _[Follow-up for #1981]_
Both elements are unnecessary wrappers around <tt>\<tr-ui-u-scalar-span\></tt>. I propose doing the following:
* Replace <tt>\<tr-ui-u-__time-stamp__-span\></tt> with <tt>\<tr-ui-u-__scalar__-span __unit="timeStampInMs"__\></tt> and
* Replace <tt>tr.ui.units.create**TimeStamp**Span(x)</tt> with <tt>tr.ui.units.create**Scalar**Span(__new tr.b.u.Scalar__(x, __tr.b.u.Unit.byName.timeStampInMs__))</tt>.
@natduca: Sounds good?
|
non_process
|
get rid of and both elements are unnecessary wrappers around i propose doing the following replace with and replace tr ui units create timestamp span x with tr ui units create scalar span new tr b u scalar x tr b u unit byname timestampinms natduca sounds good
| 0
|
4,040
| 6,972,783,515
|
IssuesEvent
|
2017-12-11 18:10:45
|
triplea-game/triplea
|
https://api.github.com/repos/triplea-game/triplea
|
reopened
|
Move install4j bundled JREs from GitHub to Linode
|
category: dev & admin process discussion type: process
|
We bundle JREs with our installer for users that do not have a Java 8 JRE installed on their machine. We currently host these bundled JREs in the triplea-game/assets repo, and the installer downloads them, if needed, directly from GitHub.
There are a few problems hosting these files on GitHub:
* We can't just provide the "latest" JRE because older installers are built with a hard-coded link to a specific path in the repo. As long as those older builds are considered compatible, we have to keep multiple JRE versions on `HEAD` in the repo (we currently have two versions for three platforms). Each JRE is approximately 350 MiB.
* Adding additional JREs increases the time to run the Gradle build on Travis due to the additional download requirements.
* Adding additional JREs increases the size of the triplea-game/assets repo significantly. Even when we can remove an older bundled JRE from `HEAD`, it stays in the repo history unless we rewrite history.
A static file server is probably a better place to host these resources rather than using a Git repo. **The purpose of this issue is to discuss possibly moving the bundled JREs to one of our Linode servers and serve them from there using Nginx (or something equivalent).**
Some issues with self-hosting include:
1. Increased bandwidth possibly leading to an increased Linode bill. It would be great if we could get some metrics from GitHub to see how often the bundled JREs are downloaded so we can predict how much traffic we'll see.
1. The host should be available as much as possible. It should not be taken down except for maintenance.
1. The host name should not change over time. We won't be able to go back and modify installers from older releases to point them to a new host. I'm not familiar with this aspect of our Linode setup, so if changing host names is common, we might have to consider having a reverse proxy or something to redirect from the old name to the new name.
@DanVanAtta @prastle @RoiEXLab @ron-murhammer Thoughts?
|
2.0
|
Move install4j bundled JREs from GitHub to Linode - We bundle JREs with our installer for users that do not have a Java 8 JRE installed on their machine. We currently host these bundled JREs in the triplea-game/assets repo, and the installer downloads them, if needed, directly from GitHub.
There are a few problems hosting these files on GitHub:
* We can't just provide the "latest" JRE because older installers are built with a hard-coded link to a specific path in the repo. As long as those older builds are considered compatible, we have to keep multiple JRE versions on `HEAD` in the repo (we currently have two versions for three platforms). Each JRE is approximately 350 MiB.
* Adding additional JREs increases the time to run the Gradle build on Travis due to the additional download requirements.
* Adding additional JREs increases the size of the triplea-game/assets repo significantly. Even when we can remove an older bundled JRE from `HEAD`, it stays in the repo history unless we rewrite history.
A static file server is probably a better place to host these resources rather than using a Git repo. **The purpose of this issue is to discuss possibly moving the bundled JREs to one of our Linode servers and serve them from there using Nginx (or something equivalent).**
Some issues with self-hosting include:
1. Increased bandwidth possibly leading to an increased Linode bill. It would be great if we could get some metrics from GitHub to see how often the bundled JREs are downloaded so we can predict how much traffic we'll see.
1. The host should be available as much as possible. It should not be taken down except for maintenance.
1. The host name should not change over time. We won't be able to go back and modify installers from older releases to point them to a new host. I'm not familiar with this aspect of our Linode setup, so if changing host names is common, we might have to consider having a reverse proxy or something to redirect from the old name to the new name.
@DanVanAtta @prastle @RoiEXLab @ron-murhammer Thoughts?
|
process
|
move bundled jres from github to linode we bundle jres with our installer for users that do not have a java jre installed on their machine we currently host these bundled jres in the triplea game assets repo and the installer downloads them if needed directly from github there are a few problems hosting these files on github we can t just provide the latest jre because older installers are built with a hard coded link to a specific path in the repo as long as those older builds are considered compatible we have to keep multiple jre versions on head in the repo we currently have two versions for three platforms each jre is approximately mib adding additional jres increases the time to run the gradle build on travis due to the additional download requirements adding additional jres increases the size of the triplea game assets repo significantly even when we can remove an older bundled jre from head it stays in the repo history unless we rewrite history a static file server is probably a better place to host these resources rather than using a git repo the purpose of this issue is to discuss possibly moving the bundled jres to one of our linode servers and serve them from there using nginx or something equivalent some issues with self hosting include increased bandwidth possibly leading to an increased linode bill it would be great if we could get some metrics from github to see how often the bundled jres are downloaded so we can predict how much traffic we ll see the host should be available as much as possible it should not be taken down except for maintenance the host name should not change over time we won t be able to go back and modify installers from older releases to point them to a new host i m not familiar with this aspect of our linode setup so if changing host names is common we might have to consider having a reverse proxy or something to redirect from the old name to the new name danvanatta prastle roiexlab ron murhammer thoughts
| 1
|
2,775
| 5,712,694,876
|
IssuesEvent
|
2017-04-19 04:46:45
|
kerubistan/kerub
|
https://api.github.com/repos/kerubistan/kerub
|
opened
|
nulls injected through json
|
bug component: security component:data processing priority: high
|
while nulls are not expected on the server, they can be injected through json
https://twitter.com/kozka/status/854439819216396288
this would be great if jackson could handle this, but it is "wontfix" https://github.com/FasterXML/jackson-module-kotlin/issues/27
|
1.0
|
nulls injected through json - while nulls are not expected on the server, they can be injected through json
https://twitter.com/kozka/status/854439819216396288
this would be great if jackson could handle this, but it is "wontfix" https://github.com/FasterXML/jackson-module-kotlin/issues/27
|
process
|
nulls injected through json while nulls are not expected on the server they can be injected through json this would be great if jackson could handle this but it is wontfix
| 1
|
22,218
| 30,768,981,159
|
IssuesEvent
|
2023-07-30 17:06:41
|
km4ack/73Linux
|
https://api.github.com/repos/km4ack/73Linux
|
closed
|
73Linux/x86LMint Update Issues
|
in process
|
So, earlier I messaged about problems with getting the IC-705 to work with WSJT-X. The fix appeared to be to update WSJT-X to the latest revision, which uses an updated HAMLIB that does appear to support the IC-705.
Unfortunately, noticing that I was unable to update WSJT-X (using the 73Linux update tool), I tried to update several other installed applications one at a time, including CHIRP, HAMLIB and FLRIG, without success. The update operation completed to the "REBOOT" dialog and upon reboot the requested application was not updated. There was no notice-able error presented in either the on-screen log or the stored on in ~/73linux/cache/logs, but no update.
Gateway x86-64 laptop with LMint installed under 73Linux; plenty of storage and memory.
Any thoughts?
Thanks,
--
Jeff Marden
N1JCM
|
1.0
|
73Linux/x86LMint Update Issues - So, earlier I messaged about problems with getting the IC-705 to work with WSJT-X. The fix appeared to be to update WSJT-X to the latest revision, which uses an updated HAMLIB that does appear to support the IC-705.
Unfortunately, noticing that I was unable to update WSJT-X (using the 73Linux update tool), I tried to update several other installed applications one at a time, including CHIRP, HAMLIB and FLRIG, without success. The update operation completed to the "REBOOT" dialog and upon reboot the requested application was not updated. There was no notice-able error presented in either the on-screen log or the stored on in ~/73linux/cache/logs, but no update.
Gateway x86-64 laptop with LMint installed under 73Linux; plenty of storage and memory.
Any thoughts?
Thanks,
--
Jeff Marden
N1JCM
|
process
|
update issues so earlier i messaged about problems with getting the ic to work with wsjt x the fix appeared to be to update wsjt x to the latest revision which uses an updated hamlib that does appear to support the ic unfortunately noticing that i was unable to update wsjt x using the update tool i tried to update several other installed applications one at a time including chirp hamlib and flrig without success the update operation completed to the reboot dialog and upon reboot the requested application was not updated there was no notice able error presented in either the on screen log or the stored on in cache logs but no update gateway laptop with lmint installed under plenty of storage and memory any thoughts thanks jeff marden
| 1
|
220,229
| 16,902,865,065
|
IssuesEvent
|
2021-06-24 00:58:54
|
bethlakshmi/gbe-divio-djangocms-python2.7
|
https://api.github.com/repos/bethlakshmi/gbe-divio-djangocms-python2.7
|
opened
|
Wrong URL in ticketing page doc
|
bug documentation
|
If you go to edit a ticket item there's some help test next to the field where you can indicate which "simple Icon" is displayed. The URL given is wrong and should be https://simplelineicons.github.io/
|
1.0
|
Wrong URL in ticketing page doc - If you go to edit a ticket item there's some help test next to the field where you can indicate which "simple Icon" is displayed. The URL given is wrong and should be https://simplelineicons.github.io/
|
non_process
|
wrong url in ticketing page doc if you go to edit a ticket item there s some help test next to the field where you can indicate which simple icon is displayed the url given is wrong and should be
| 0
|
251,129
| 8,000,370,873
|
IssuesEvent
|
2018-07-22 15:11:18
|
krshubham/interview-prep
|
https://api.github.com/repos/krshubham/interview-prep
|
closed
|
App Tooling Upgrade
|
Priority:low
|
## DB
Move to RethinkDB over MongoDB. It's perfect for our usage. https://rethinkdb.com/faq/
## Auth
Use Passport.js (when we move to the model where users can sign in and reply)
## API
Move to GraphQl (Apollo toolchain)
##Express
Setup it like it's here: https://github.com/withspectrum/spectrum/blob/alpha/api/index.js
## Javascript Tooling
There is a lot of work that's been done in the JS ecosystem to finally put the fires down between developers. Why not use them?
Our team is primarily using VSCode as the editor, but feel free to use any which supports these tools (most popular editor have extensions for them):
* Prettier.
* Eslint
* Flow (will introduce when code has to scale)
Eslint
* Too many console.logs spoil the broth. Warning for console.logs
* Introduce Airbnb's eslint config.
Good extensions to have:
https://marketplace.visualstudio.com/items?itemName=burkeholland.simple-react-snippets
|
1.0
|
App Tooling Upgrade -
## DB
Move to RethinkDB over MongoDB. It's perfect for our usage. https://rethinkdb.com/faq/
## Auth
Use Passport.js (when we move to the model where users can sign in and reply)
## API
Move to GraphQl (Apollo toolchain)
##Express
Setup it like it's here: https://github.com/withspectrum/spectrum/blob/alpha/api/index.js
## Javascript Tooling
There is a lot of work that's been done in the JS ecosystem to finally put the fires down between developers. Why not use them?
Our team is primarily using VSCode as the editor, but feel free to use any which supports these tools (most popular editor have extensions for them):
* Prettier.
* Eslint
* Flow (will introduce when code has to scale)
Eslint
* Too many console.logs spoil the broth. Warning for console.logs
* Introduce Airbnb's eslint config.
Good extensions to have:
https://marketplace.visualstudio.com/items?itemName=burkeholland.simple-react-snippets
|
non_process
|
app tooling upgrade db move to rethinkdb over mongodb it s perfect for our usage auth use passport js when we move to the model where users can sign in and reply api move to graphql apollo toolchain express setup it like it s here javascript tooling there is a lot of work that s been done in the js ecosystem to finally put the fires down between developers why not use them our team is primarily using vscode as the editor but feel free to use any which supports these tools most popular editor have extensions for them prettier eslint flow will introduce when code has to scale eslint too many console logs spoil the broth warning for console logs introduce airbnb s eslint config good extensions to have
| 0
|
22,880
| 20,424,097,350
|
IssuesEvent
|
2022-02-24 00:40:19
|
bevyengine/bevy
|
https://api.github.com/repos/bevyengine/bevy
|
closed
|
#[derive(Query)]
|
C-Enhancement A-ECS C-Usability
|
**What problem does this solve or what need does it fill?**
Giant tuples within query (`Query<(&Foo, &Bar, &mut Baz)>`) is really annoying as then everything has to be referenced by `.0`, etc, and if I decide to add `Entity` to the list I have to shift everything.
**Describe the solution would you like?**
Make a derive macro `#[derive(Query)]` which implements `HecsQuery` for the structure in question; as such the structure can be used as a query.
**Describe the alternative(s) you've considered?**
None.
**Additional context**
This could also support `Or` by allowing deriving an enum.
|
True
|
#[derive(Query)] - **What problem does this solve or what need does it fill?**
Giant tuples within query (`Query<(&Foo, &Bar, &mut Baz)>`) is really annoying as then everything has to be referenced by `.0`, etc, and if I decide to add `Entity` to the list I have to shift everything.
**Describe the solution would you like?**
Make a derive macro `#[derive(Query)]` which implements `HecsQuery` for the structure in question; as such the structure can be used as a query.
**Describe the alternative(s) you've considered?**
None.
**Additional context**
This could also support `Or` by allowing deriving an enum.
|
non_process
|
what problem does this solve or what need does it fill giant tuples within query query is really annoying as then everything has to be referenced by etc and if i decide to add entity to the list i have to shift everything describe the solution would you like make a derive macro which implements hecsquery for the structure in question as such the structure can be used as a query describe the alternative s you ve considered none additional context this could also support or by allowing deriving an enum
| 0
|
20,687
| 27,358,213,024
|
IssuesEvent
|
2023-02-27 14:16:31
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
Is the listed Throw syntax correct?
|
automation/svc triaged cxp doc-enhancement process-automation/subsvc Pri2
|
[Enter feedback here]
I think the Throw syntax in the screenshot in learn is incorrect.
Is this correct?

---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: 8ab9bff0-9179-827d-1a8d-c58a98625070
* Version Independent ID: bf46f240-6c29-f040-6b0f-b454606b04a1
* Content: [Handle errors in Azure Automation graphical runbooks](https://learn.microsoft.com/en-us/azure/automation/automation-runbook-graphical-error-handling#turn-exceptions-into-non-terminating-errors)
* Content Source: [articles/automation/automation-runbook-graphical-error-handling.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/automation/automation-runbook-graphical-error-handling.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @SnehaSudhirG
* Microsoft Alias: **sudhirsneha**
|
1.0
|
Is the listed Throw syntax correct? -
[Enter feedback here]
I think the Throw syntax in the screenshot in learn is incorrect.
Is this correct?

---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: 8ab9bff0-9179-827d-1a8d-c58a98625070
* Version Independent ID: bf46f240-6c29-f040-6b0f-b454606b04a1
* Content: [Handle errors in Azure Automation graphical runbooks](https://learn.microsoft.com/en-us/azure/automation/automation-runbook-graphical-error-handling#turn-exceptions-into-non-terminating-errors)
* Content Source: [articles/automation/automation-runbook-graphical-error-handling.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/automation/automation-runbook-graphical-error-handling.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @SnehaSudhirG
* Microsoft Alias: **sudhirsneha**
|
process
|
is the listed throw syntax correct i think the throw syntax in the screenshot in learn is incorrect is this correct document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login snehasudhirg microsoft alias sudhirsneha
| 1
|
12,054
| 14,739,179,187
|
IssuesEvent
|
2021-01-07 06:39:48
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
'No records found' should be displayed when admin doesn't have any site level, study level, app level permission
|
Bug P1 Participant manager Process: Fixed Process: Tested QA Process: Tested dev
|
1. 'No records found' text should be displayed when there are no data in the database
2. 'No records found' text should be displayed when there are some data in the database and the user does not have any
permission for sites
3. Proper custom messages should be displayed when a normal admin user doesn't meet conditions related to permission for the following
i. Study level permission
AR : 'This view displays study-wise enrollment if you manage multiple sites' error message is displayed
ER : 'This view displays study-wise enrollment if you manage multiple sites' text should be displayed in the attached screen instead of an error message
ii. App-level permission
AR : 'This view displays app-wise enrollment if you manage multiple studies.' error message is displayed
ER : 'This view displays app-wise enrollment if you manage multiple studies' text should be displayed in the attached screen instead of an error message
ER:

|
3.0
|
'No records found' should be displayed when admin doesn't have any site level, study level, app level permission - 1. 'No records found' text should be displayed when there are no data in the database
2. 'No records found' text should be displayed when there are some data in the database and the user does not have any
permission for sites
3. Proper custom messages should be displayed when a normal admin user doesn't meet conditions related to permission for the following
i. Study level permission
AR : 'This view displays study-wise enrollment if you manage multiple sites' error message is displayed
ER : 'This view displays study-wise enrollment if you manage multiple sites' text should be displayed in the attached screen instead of an error message
ii. App-level permission
AR : 'This view displays app-wise enrollment if you manage multiple studies.' error message is displayed
ER : 'This view displays app-wise enrollment if you manage multiple studies' text should be displayed in the attached screen instead of an error message
ER:

|
process
|
no records found should be displayed when admin doesn t have any site level study level app level permission no records found text should be displayed when there are no data in the database no records found text should be displayed when there are some data in the database and the user does not have any permission for sites proper custom messages should be displayed when a normal admin user doesn t meet conditions related to permission for the following i study level permission ar this view displays study wise enrollment if you manage multiple sites error message is displayed er this view displays study wise enrollment if you manage multiple sites text should be displayed in the attached screen instead of an error message ii app level permission ar this view displays app wise enrollment if you manage multiple studies error message is displayed er this view displays app wise enrollment if you manage multiple studies text should be displayed in the attached screen instead of an error message er
| 1
|
9,162
| 3,258,419,954
|
IssuesEvent
|
2015-10-20 22:15:35
|
radical-cybertools/radical.pilot
|
https://api.github.com/repos/radical-cybertools/radical.pilot
|
closed
|
a better title?
|
documentation
|
4.2 "Obtaining Unit Details" --> 4.2 "Inspecting Execution" or something to that effect?
|
1.0
|
a better title? - 4.2 "Obtaining Unit Details" --> 4.2 "Inspecting Execution" or something to that effect?
|
non_process
|
a better title obtaining unit details inspecting execution or something to that effect
| 0
|
19,470
| 25,767,606,786
|
IssuesEvent
|
2022-12-09 04:07:00
|
dtcenter/MET
|
https://api.github.com/repos/dtcenter/MET
|
closed
|
Fix logic in reading AERONET v3 data
|
type: bug component: user support requestor: NOAA/EMC reporting: DTC NOAA R2O required: FOR OFFICIAL RELEASE MET: PreProcessing Tools (Point) priority: high
|
## Describe the Problem ##
This issue arose with the METplus Discussion [dtcenter/METplus#1888](https://github.com/dtcenter/METplus/discussions/1888). When the user obtained AERONET version 3 data from the [official site](https://aeronet.gsfc.nasa.gov) it was determined that the format was not consistent with the format of the data we currently use for our [unit testing](https://dtcenter.ucar.edu/dfiles/code/METplus/MET/MET_unit_test/unit_test/obs_data/aeronet/20150917_20150926_Table_Mountain.lev20). Sample files from the new format and @hsoh-u's analysis of what would need to be done for this issue can be viewed within the Discussion at this [post](https://github.com/dtcenter/METplus/discussions/1888#discussioncomment-4319771).
### Expected Behavior ###
The MET code should be able to read AERONET v3 data obtained from the official site.
### Environment ###
Describe your runtime environment:
*1. Machine: not specific*
*2. OS: not specific*
*3. Software version number: all current releases*
### To Reproduce ###
Describe the steps to reproduce the behavior:
*1. Go to '...'*
*2. Click on '....'*
*3. Scroll down to '....'*
*4. See error*
*Post relevant sample data following these instructions:*
*https://dtcenter.org/community-code/model-evaluation-tools-met/met-help-desk#ftp*
Review the log file from Partha and read the data files he provided.
### Relevant Deadlines ###
MET-11.0.0 Official Release (12/7 - 12/9)
### Funding Source ###
2773542
## Define the Metadata ##
### Assignee ###
- [x] Select **engineer(s)** or **no engineer** required
- [ ] Select **scientist(s)** or **no scientist** required
### Labels ###
- [x] Select **component(s)**
- [x] Select **priority**
- [x] Select **requestor(s)**
### Projects and Milestone ###
- [x] Select **Organization** level **Project** for support of the current coordinated release
- [x] Select **Repository** level **Project** for development toward the next official release or add **alert: NEED PROJECT ASSIGNMENT** label
- [x] Select **Milestone** as the next bugfix version
## Define Related Issue(s) ##
Consider the impact to the other METplus components.
- [x] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdataio](https://github.com/dtcenter/METdataio/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose)
## Bugfix Checklist ##
See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details.
- [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**.
- [ ] Fork this repository or create a branch of **main_\<Version>**.
Branch name: `bugfix_<Issue Number>_main_<Version>_<Description>`
- [ ] Fix the bug and test your changes.
- [ ] Add/update log messages for easier debugging.
- [ ] Add/update unit tests.
- [ ] Add/update documentation.
- [ ] Push local changes to GitHub.
- [ ] Submit a pull request to merge into **main_\<Version>**.
Pull request: `bugfix <Issue Number> main_<Version> <Description>`
- [ ] Define the pull request metadata, as permissions allow.
Select: **Reviewer(s)** and **Linked issues**
Select: **Organization** level software support **Project** for the current coordinated release
Select: **Milestone** as the next bugfix version
- [ ] Iterate until the reviewer(s) accept and merge your changes.
- [ ] Delete your fork or branch.
- [ ] Complete the steps above to fix the bug on the **develop** branch.
Branch name: `bugfix_<Issue Number>_develop_<Description>`
Pull request: `bugfix <Issue Number> develop <Description>`
Select: **Reviewer(s)** and **Linked issues**
Select: **Repository** level development cycle **Project** for the next official release
Select: **Milestone** as the next official version
- [ ] Close this issue.
|
1.0
|
Fix logic in reading AERONET v3 data - ## Describe the Problem ##
This issue arose with the METplus Discussion [dtcenter/METplus#1888](https://github.com/dtcenter/METplus/discussions/1888). When the user obtained AERONET version 3 data from the [official site](https://aeronet.gsfc.nasa.gov) it was determined that the format was not consistent with the format of the data we currently use for our [unit testing](https://dtcenter.ucar.edu/dfiles/code/METplus/MET/MET_unit_test/unit_test/obs_data/aeronet/20150917_20150926_Table_Mountain.lev20). Sample files from the new format and @hsoh-u's analysis of what would need to be done for this issue can be viewed within the Discussion at this [post](https://github.com/dtcenter/METplus/discussions/1888#discussioncomment-4319771).
### Expected Behavior ###
The MET code should be able to read AERONET v3 data obtained from the official site.
### Environment ###
Describe your runtime environment:
*1. Machine: not specific*
*2. OS: not specific*
*3. Software version number: all current releases*
### To Reproduce ###
Describe the steps to reproduce the behavior:
*1. Go to '...'*
*2. Click on '....'*
*3. Scroll down to '....'*
*4. See error*
*Post relevant sample data following these instructions:*
*https://dtcenter.org/community-code/model-evaluation-tools-met/met-help-desk#ftp*
Review the log file from Partha and read the data files he provided.
### Relevant Deadlines ###
MET-11.0.0 Official Release (12/7 - 12/9)
### Funding Source ###
2773542
## Define the Metadata ##
### Assignee ###
- [x] Select **engineer(s)** or **no engineer** required
- [ ] Select **scientist(s)** or **no scientist** required
### Labels ###
- [x] Select **component(s)**
- [x] Select **priority**
- [x] Select **requestor(s)**
### Projects and Milestone ###
- [x] Select **Organization** level **Project** for support of the current coordinated release
- [x] Select **Repository** level **Project** for development toward the next official release or add **alert: NEED PROJECT ASSIGNMENT** label
- [x] Select **Milestone** as the next bugfix version
## Define Related Issue(s) ##
Consider the impact to the other METplus components.
- [x] [METplus](https://github.com/dtcenter/METplus/issues/new/choose), [MET](https://github.com/dtcenter/MET/issues/new/choose), [METdataio](https://github.com/dtcenter/METdataio/issues/new/choose), [METviewer](https://github.com/dtcenter/METviewer/issues/new/choose), [METexpress](https://github.com/dtcenter/METexpress/issues/new/choose), [METcalcpy](https://github.com/dtcenter/METcalcpy/issues/new/choose), [METplotpy](https://github.com/dtcenter/METplotpy/issues/new/choose)
## Bugfix Checklist ##
See the [METplus Workflow](https://metplus.readthedocs.io/en/latest/Contributors_Guide/github_workflow.html) for details.
- [ ] Complete the issue definition above, including the **Time Estimate** and **Funding Source**.
- [ ] Fork this repository or create a branch of **main_\<Version>**.
Branch name: `bugfix_<Issue Number>_main_<Version>_<Description>`
- [ ] Fix the bug and test your changes.
- [ ] Add/update log messages for easier debugging.
- [ ] Add/update unit tests.
- [ ] Add/update documentation.
- [ ] Push local changes to GitHub.
- [ ] Submit a pull request to merge into **main_\<Version>**.
Pull request: `bugfix <Issue Number> main_<Version> <Description>`
- [ ] Define the pull request metadata, as permissions allow.
Select: **Reviewer(s)** and **Linked issues**
Select: **Organization** level software support **Project** for the current coordinated release
Select: **Milestone** as the next bugfix version
- [ ] Iterate until the reviewer(s) accept and merge your changes.
- [ ] Delete your fork or branch.
- [ ] Complete the steps above to fix the bug on the **develop** branch.
Branch name: `bugfix_<Issue Number>_develop_<Description>`
Pull request: `bugfix <Issue Number> develop <Description>`
Select: **Reviewer(s)** and **Linked issues**
Select: **Repository** level development cycle **Project** for the next official release
Select: **Milestone** as the next official version
- [ ] Close this issue.
|
process
|
fix logic in reading aeronet data describe the problem this issue arose with the metplus discussion when the user obtained aeronet version data from the it was determined that the format was not consistent with the format of the data we currently use for our sample files from the new format and hsoh u s analysis of what would need to be done for this issue can be viewed within the discussion at this expected behavior the met code should be able to read aeronet data obtained from the official site environment describe your runtime environment machine not specific os not specific software version number all current releases to reproduce describe the steps to reproduce the behavior go to click on scroll down to see error post relevant sample data following these instructions review the log file from partha and read the data files he provided relevant deadlines met official release funding source define the metadata assignee select engineer s or no engineer required select scientist s or no scientist required labels select component s select priority select requestor s projects and milestone select organization level project for support of the current coordinated release select repository level project for development toward the next official release or add alert need project assignment label select milestone as the next bugfix version define related issue s consider the impact to the other metplus components bugfix checklist see the for details complete the issue definition above including the time estimate and funding source fork this repository or create a branch of main branch name bugfix main fix the bug and test your changes add update log messages for easier debugging add update unit tests add update documentation push local changes to github submit a pull request to merge into main pull request bugfix main define the pull request metadata as permissions allow select reviewer s and linked issues select organization level software support project for the current coordinated release select milestone as the next bugfix version iterate until the reviewer s accept and merge your changes delete your fork or branch complete the steps above to fix the bug on the develop branch branch name bugfix develop pull request bugfix develop select reviewer s and linked issues select repository level development cycle project for the next official release select milestone as the next official version close this issue
| 1
|
13,135
| 15,555,305,893
|
IssuesEvent
|
2021-03-16 05:51:59
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
gdal_rasterize update
|
Feature Request Processing
|
Hello everyone,
according to gdal_rasterize documentation it is possible to use -3d option in order to use the Z attribute to rasterize. Is it possible to include that feature in QGIS processing?
I tried, as a test, to include the "-3d" in advanced parameters, but an error is thrown because the parameter "burn value" (claimed as optional) is anyway called in the GDAL string (while the filed attribute can be skipped):
```
gdal_rasterize -l INPUT -burn 0.0 -tr 5.0 5.0 -a_nodata 0.0 -te 2309012.5 4638817.5 2312552.5 4641697.5 -ot Float32 -of GTiff -co COMPRESS=DEFLATE -co PREDICTOR=2 -co ZLEVEL=9 -3d C:/Users/dtalledo/AppData/Local/Temp/processing_jPHwjv/856fc12d75c345d780105fcae4220de7/INPUT.gpkg C:/Users/dtalledo/AppData/Local/Temp/processing_jPHwjv/90b067e3959643b9b1093a3eb5c8666f/OUTPUT.tif
Risultato comando GDAL:
ERROR 6: One and only one of -3d, -burn or -a is required.
Usage: gdal_rasterize [-b band]* [-i] [-at]
{[-burn value]* | [-a attribute_name] | [-3d]} [-add]
[-l layername]* [-where expression] [-sql select_statement]
[-dialect dialect] [-of format] [-a_srs srs_def] [-to "NAME=VALUE"]*
[-co "NAME=VALUE"]* [-a_nodata value] [-init value]*
[-te xmin ymin xmax ymax] [-tr xres yres] [-tap] [-ts width height]
[-ot {Byte/Int16/UInt16/UInt32/Int32/Float32/Float64/
CInt16/CInt32/CFloat32/CFloat64}] [-optim {[AUTO]/VECTOR/RASTER}] [-q]
<src_datasource> <dst_filename>
Il processo ha restituito un codice di errore 1
```
|
1.0
|
gdal_rasterize update - Hello everyone,
according to gdal_rasterize documentation it is possible to use -3d option in order to use the Z attribute to rasterize. Is it possible to include that feature in QGIS processing?
I tried, as a test, to include the "-3d" in advanced parameters, but an error is thrown because the parameter "burn value" (claimed as optional) is anyway called in the GDAL string (while the filed attribute can be skipped):
```
gdal_rasterize -l INPUT -burn 0.0 -tr 5.0 5.0 -a_nodata 0.0 -te 2309012.5 4638817.5 2312552.5 4641697.5 -ot Float32 -of GTiff -co COMPRESS=DEFLATE -co PREDICTOR=2 -co ZLEVEL=9 -3d C:/Users/dtalledo/AppData/Local/Temp/processing_jPHwjv/856fc12d75c345d780105fcae4220de7/INPUT.gpkg C:/Users/dtalledo/AppData/Local/Temp/processing_jPHwjv/90b067e3959643b9b1093a3eb5c8666f/OUTPUT.tif
Risultato comando GDAL:
ERROR 6: One and only one of -3d, -burn or -a is required.
Usage: gdal_rasterize [-b band]* [-i] [-at]
{[-burn value]* | [-a attribute_name] | [-3d]} [-add]
[-l layername]* [-where expression] [-sql select_statement]
[-dialect dialect] [-of format] [-a_srs srs_def] [-to "NAME=VALUE"]*
[-co "NAME=VALUE"]* [-a_nodata value] [-init value]*
[-te xmin ymin xmax ymax] [-tr xres yres] [-tap] [-ts width height]
[-ot {Byte/Int16/UInt16/UInt32/Int32/Float32/Float64/
CInt16/CInt32/CFloat32/CFloat64}] [-optim {[AUTO]/VECTOR/RASTER}] [-q]
<src_datasource> <dst_filename>
Il processo ha restituito un codice di errore 1
```
|
process
|
gdal rasterize update hello everyone according to gdal rasterize documentation it is possible to use option in order to use the z attribute to rasterize is it possible to include that feature in qgis processing i tried as a test to include the in advanced parameters but an error is thrown because the parameter burn value claimed as optional is anyway called in the gdal string while the filed attribute can be skipped gdal rasterize l input burn tr a nodata te ot of gtiff co compress deflate co predictor co zlevel c users dtalledo appdata local temp processing jphwjv input gpkg c users dtalledo appdata local temp processing jphwjv output tif risultato comando gdal error one and only one of burn or a is required usage gdal rasterize ot byte vector raster il processo ha restituito un codice di errore
| 1
|
4,502
| 7,348,891,265
|
IssuesEvent
|
2018-03-08 08:42:30
|
qgis/QGIS-Documentation
|
https://api.github.com/repos/qgis/QGIS-Documentation
|
opened
|
Add more relation between algorithms help
|
Easy Processing
|
It could be nice to have a more intensive use of the "See Also" section of alg description and link them.
For example, a minima, when alg A is listed in alg B see also, the inverse should be true. Which is not the case currently. But there might also be missing links.
|
1.0
|
Add more relation between algorithms help - It could be nice to have a more intensive use of the "See Also" section of alg description and link them.
For example, a minima, when alg A is listed in alg B see also, the inverse should be true. Which is not the case currently. But there might also be missing links.
|
process
|
add more relation between algorithms help it could be nice to have a more intensive use of the see also section of alg description and link them for example a minima when alg a is listed in alg b see also the inverse should be true which is not the case currently but there might also be missing links
| 1
|
287,160
| 8,805,120,492
|
IssuesEvent
|
2018-12-26 17:36:01
|
strapi/strapi
|
https://api.github.com/repos/strapi/strapi
|
closed
|
Add support for GraphQL Apollo server tracing
|
pr: 🚀 New feature priority: low
|
- [ ] **I have created my request on the Product Board before I submitted this issue**
- [x] **I have looked at all the other requests on the Product Board before I submitted this issue**
(Feature Request not submitted to Product board as I plan to start playing with adding this via a PR myself.)
**Please describe your feature request:**
Per the following:
https://github.com/apollographql/apollo-server/tree/master/packages/apollo-tracing
https://github.com/apollographql/apollo-tracing
I'm requesting that the Apollo server tracing be enabled so that we can see live performance metrics of a GraphQL based query to determine load and plan out/negate potential damaging query and mutation requests.
|
1.0
|
Add support for GraphQL Apollo server tracing - - [ ] **I have created my request on the Product Board before I submitted this issue**
- [x] **I have looked at all the other requests on the Product Board before I submitted this issue**
(Feature Request not submitted to Product board as I plan to start playing with adding this via a PR myself.)
**Please describe your feature request:**
Per the following:
https://github.com/apollographql/apollo-server/tree/master/packages/apollo-tracing
https://github.com/apollographql/apollo-tracing
I'm requesting that the Apollo server tracing be enabled so that we can see live performance metrics of a GraphQL based query to determine load and plan out/negate potential damaging query and mutation requests.
|
non_process
|
add support for graphql apollo server tracing i have created my request on the product board before i submitted this issue i have looked at all the other requests on the product board before i submitted this issue feature request not submitted to product board as i plan to start playing with adding this via a pr myself please describe your feature request per the following i m requesting that the apollo server tracing be enabled so that we can see live performance metrics of a graphql based query to determine load and plan out negate potential damaging query and mutation requests
| 0
|
41
| 2,507,672,152
|
IssuesEvent
|
2015-01-12 19:57:50
|
tinkerpop/tinkerpop3
|
https://api.github.com/repos/tinkerpop/tinkerpop3
|
closed
|
[Proposal] has() as a step modulator.
|
enhancement process
|
Here is a big internal change that may be useful, may not.
A `Step` implements `HasContainerHolder` if it stores `HasContainers`. E.g. `HasStep`, `TinkerGraphStep`, `Neo4jGraphStep`. Why not just make it such that:
```java
public GraphTraversal<S,E> has(...) {
if(previousStep instanceof HasContainer)
previousStep.addHasContainer(new HasContainer(...));
else
this.addStep(new HasStep(new HasContainer(...)));
return this;
}
```
Why is this cool?
. GraphStep can implement `HasContainerHolder`.
. There is no need for a `TinkerGraphStepStrategy`, `Neo4jGraphStepStrategy`, they simply implement `HasContainerHolder`.
. `RepeatStep` can implement `HasContainerHolder`. ... see below why this is cool.
```java
g.V().repeat(__.out()).until().has(label,'person')
```
In other words, repeat `out()` until you reach a person.
@mbroecheler @dkuppitz @pietermartin @spmallette
|
1.0
|
[Proposal] has() as a step modulator. - Here is a big internal change that may be useful, may not.
A `Step` implements `HasContainerHolder` if it stores `HasContainers`. E.g. `HasStep`, `TinkerGraphStep`, `Neo4jGraphStep`. Why not just make it such that:
```java
public GraphTraversal<S,E> has(...) {
if(previousStep instanceof HasContainer)
previousStep.addHasContainer(new HasContainer(...));
else
this.addStep(new HasStep(new HasContainer(...)));
return this;
}
```
Why is this cool?
. GraphStep can implement `HasContainerHolder`.
. There is no need for a `TinkerGraphStepStrategy`, `Neo4jGraphStepStrategy`, they simply implement `HasContainerHolder`.
. `RepeatStep` can implement `HasContainerHolder`. ... see below why this is cool.
```java
g.V().repeat(__.out()).until().has(label,'person')
```
In other words, repeat `out()` until you reach a person.
@mbroecheler @dkuppitz @pietermartin @spmallette
|
process
|
has as a step modulator here is a big internal change that may be useful may not a step implements hascontainerholder if it stores hascontainers e g hasstep tinkergraphstep why not just make it such that java public graphtraversal has if previousstep instanceof hascontainer previousstep addhascontainer new hascontainer else this addstep new hasstep new hascontainer return this why is this cool graphstep can implement hascontainerholder there is no need for a tinkergraphstepstrategy they simply implement hascontainerholder repeatstep can implement hascontainerholder see below why this is cool java g v repeat out until has label person in other words repeat out until you reach a person mbroecheler dkuppitz pietermartin spmallette
| 1
|
11,291
| 14,100,076,964
|
IssuesEvent
|
2020-11-06 03:08:34
|
dita-ot/dita-ot
|
https://api.github.com/repos/dita-ot/dita-ot
|
closed
|
Retain conref and keyref information after preprocessing
|
feature good first issue preprocess preprocess/conref preprocess/keyref priority/medium stale
|
Add namespaced attributes to retain conref and keyref source information. This will allow advanced users to create e.g. review PDFs which show the source of each block.
The feature should be controlled by dynamic configuration; by default the feature should be disabled.
|
3.0
|
Retain conref and keyref information after preprocessing - Add namespaced attributes to retain conref and keyref source information. This will allow advanced users to create e.g. review PDFs which show the source of each block.
The feature should be controlled by dynamic configuration; by default the feature should be disabled.
|
process
|
retain conref and keyref information after preprocessing add namespaced attributes to retain conref and keyref source information this will allow advanced users to create e g review pdfs which show the source of each block the feature should be controlled by dynamic configuration by default the feature should be disabled
| 1
|
64,149
| 14,657,456,207
|
IssuesEvent
|
2020-12-28 15:38:13
|
fu1771695yongxie/yarn
|
https://api.github.com/repos/fu1771695yongxie/yarn
|
opened
|
CVE-2018-3750 (High) detected in io.js6ed791c665de2c1838f6080a1b377b0008cf535b
|
security vulnerability
|
## CVE-2018-3750 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>io.js6ed791c665de2c1838f6080a1b377b0008cf535b</b></p></summary>
<p>
<p>Node.js JavaScript runtime :sparkles::turtle::rocket::sparkles:</p>
<p>Library home page: <a href=https://github.com/iojs/io.js.git>https://github.com/iojs/io.js.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/yarn/commit/b0308a6bc0041ba9a7c0fefc30be7721760dfd37">b0308a6bc0041ba9a7c0fefc30be7721760dfd37</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (0)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The utilities function in all versions <= 0.5.0 of the deep-extend node module can be tricked into modifying the prototype of Object when the attacker can control part of the structure passed to this function. This can let an attacker add or modify existing properties that will exist on all objects.
<p>Publish Date: 2018-07-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3750>CVE-2018-3750</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-3750">http://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-3750</a></p>
<p>Release Date: 2019-01-24</p>
<p>Fix Resolution: 0.5.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2018-3750 (High) detected in io.js6ed791c665de2c1838f6080a1b377b0008cf535b - ## CVE-2018-3750 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>io.js6ed791c665de2c1838f6080a1b377b0008cf535b</b></p></summary>
<p>
<p>Node.js JavaScript runtime :sparkles::turtle::rocket::sparkles:</p>
<p>Library home page: <a href=https://github.com/iojs/io.js.git>https://github.com/iojs/io.js.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/yarn/commit/b0308a6bc0041ba9a7c0fefc30be7721760dfd37">b0308a6bc0041ba9a7c0fefc30be7721760dfd37</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (0)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The utilities function in all versions <= 0.5.0 of the deep-extend node module can be tricked into modifying the prototype of Object when the attacker can control part of the structure passed to this function. This can let an attacker add or modify existing properties that will exist on all objects.
<p>Publish Date: 2018-07-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3750>CVE-2018-3750</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-3750">http://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-3750</a></p>
<p>Release Date: 2019-01-24</p>
<p>Fix Resolution: 0.5.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve high detected in io cve high severity vulnerability vulnerable library io node js javascript runtime sparkles turtle rocket sparkles library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details the utilities function in all versions of the deep extend node module can be tricked into modifying the prototype of object when the attacker can control part of the structure passed to this function this can let an attacker add or modify existing properties that will exist on all objects publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
12,282
| 14,791,666,224
|
IssuesEvent
|
2021-01-12 13:48:24
|
prisma/prisma
|
https://api.github.com/repos/prisma/prisma
|
closed
|
A simple many to many relationship causes PANIC: 1
|
bug/2-confirmed kind/bug process/candidate team/client topic: broken query
|
## Bug description
Trying to implement a simple many to many relationship results in findMany causing panic
## How to reproduce
I used the following schema to create a database:
```prisma
datasource db {
provider = "postgresql"
url = "postgresql://[snipped]"
}
generator client {
provider = "prisma-client-js"
}
model Image {
hash String @id
tags Tag[] @relation(name: "tags")
}
model Tag {
name String @id
images Image[] @relation(name: "tags")
}
```
Then lanched prisma studio and created two Image instances then saved them and created a single tag, then I associated both images with that tag. After that any prisma.image.findMany query that includes or selects images fails with a panic message (including those of prisma studio).
## Expected behavior
The query should not fail since it is almost a copy of a documentation example.
```prisma
//Copied from https://www.prisma.io/docs/support/help-articles/working-with-many-to-many-relations
model Post {
id Int @id @default(autoincrement())
title String
tags Tag[]
}
model Tag {
id Int @id @default(autoincrement())
name String @unique
posts Post[]
}
```
## Prisma information
Included above
## Environment & setup
<!-- In which environment does the problem occur -->
- OS: Linux (Manjaro)
- Database: PostgreSQL
- Node.js version: 15.5.0
- Prisma version:
<!--[Run `prisma -v` to see your Prisma version and paste it between the ´´´]-->
```
@prisma/cli : 2.13.1
@prisma/client : 2.13.1
Current platform : debian-openssl-1.1.x
Query Engine : query-engine fcbc4bb2d306c86c28014f596b1e8c7980af8bd4 (at ../../../../../home/kranga/.nvm/versions/node/v15.5.0/lib/node_modules/@prisma/cli/node_modules/@prisma/engines/query-engine-debian-openssl-1.1.x)
Migration Engine : migration-engine-cli fcbc4bb2d306c86c28014f596b1e8c7980af8bd4 (at ../../../../../home/kranga/.nvm/versions/node/v15.5.0/lib/node_modules/@prisma/cli/node_modules/@prisma/engines/migration-engine-debian-openssl-1.1.x)
Introspection Engine : introspection-core fcbc4bb2d306c86c28014f596b1e8c7980af8bd4 (at ../../../../../home/kranga/.nvm/versions/node/v15.5.0/lib/node_modules/@prisma/cli/node_modules/@prisma/engines/introspection-engine-debian-openssl-1.1.x)
Format Binary : prisma-fmt fcbc4bb2d306c86c28014f596b1e8c7980af8bd4 (at ../../../../../home/kranga/.nvm/versions/node/v15.5.0/lib/node_modules/@prisma/cli/node_modules/@prisma/engines/prisma-fmt-debian-openssl-1.1.x)
Studio : 0.329.0
```
|
1.0
|
A simple many to many relationship causes PANIC: 1 - ## Bug description
Trying to implement a simple many to many relationship results in findMany causing panic
## How to reproduce
I used the following schema to create a database:
```prisma
datasource db {
provider = "postgresql"
url = "postgresql://[snipped]"
}
generator client {
provider = "prisma-client-js"
}
model Image {
hash String @id
tags Tag[] @relation(name: "tags")
}
model Tag {
name String @id
images Image[] @relation(name: "tags")
}
```
Then lanched prisma studio and created two Image instances then saved them and created a single tag, then I associated both images with that tag. After that any prisma.image.findMany query that includes or selects images fails with a panic message (including those of prisma studio).
## Expected behavior
The query should not fail since it is almost a copy of a documentation example.
```prisma
//Copied from https://www.prisma.io/docs/support/help-articles/working-with-many-to-many-relations
model Post {
id Int @id @default(autoincrement())
title String
tags Tag[]
}
model Tag {
id Int @id @default(autoincrement())
name String @unique
posts Post[]
}
```
## Prisma information
Included above
## Environment & setup
<!-- In which environment does the problem occur -->
- OS: Linux (Manjaro)
- Database: PostgreSQL
- Node.js version: 15.5.0
- Prisma version:
<!--[Run `prisma -v` to see your Prisma version and paste it between the ´´´]-->
```
@prisma/cli : 2.13.1
@prisma/client : 2.13.1
Current platform : debian-openssl-1.1.x
Query Engine : query-engine fcbc4bb2d306c86c28014f596b1e8c7980af8bd4 (at ../../../../../home/kranga/.nvm/versions/node/v15.5.0/lib/node_modules/@prisma/cli/node_modules/@prisma/engines/query-engine-debian-openssl-1.1.x)
Migration Engine : migration-engine-cli fcbc4bb2d306c86c28014f596b1e8c7980af8bd4 (at ../../../../../home/kranga/.nvm/versions/node/v15.5.0/lib/node_modules/@prisma/cli/node_modules/@prisma/engines/migration-engine-debian-openssl-1.1.x)
Introspection Engine : introspection-core fcbc4bb2d306c86c28014f596b1e8c7980af8bd4 (at ../../../../../home/kranga/.nvm/versions/node/v15.5.0/lib/node_modules/@prisma/cli/node_modules/@prisma/engines/introspection-engine-debian-openssl-1.1.x)
Format Binary : prisma-fmt fcbc4bb2d306c86c28014f596b1e8c7980af8bd4 (at ../../../../../home/kranga/.nvm/versions/node/v15.5.0/lib/node_modules/@prisma/cli/node_modules/@prisma/engines/prisma-fmt-debian-openssl-1.1.x)
Studio : 0.329.0
```
|
process
|
a simple many to many relationship causes panic bug description trying to implement a simple many to many relationship results in findmany causing panic how to reproduce i used the following schema to create a database prisma datasource db provider postgresql url postgresql generator client provider prisma client js model image hash string id tags tag relation name tags model tag name string id images image relation name tags then lanched prisma studio and created two image instances then saved them and created a single tag then i associated both images with that tag after that any prisma image findmany query that includes or selects images fails with a panic message including those of prisma studio expected behavior the query should not fail since it is almost a copy of a documentation example prisma copied from model post id int id default autoincrement title string tags tag model tag id int id default autoincrement name string unique posts post prisma information included above environment setup os linux manjaro database postgresql node js version prisma version prisma cli prisma client current platform debian openssl x query engine query engine at home kranga nvm versions node lib node modules prisma cli node modules prisma engines query engine debian openssl x migration engine migration engine cli at home kranga nvm versions node lib node modules prisma cli node modules prisma engines migration engine debian openssl x introspection engine introspection core at home kranga nvm versions node lib node modules prisma cli node modules prisma engines introspection engine debian openssl x format binary prisma fmt at home kranga nvm versions node lib node modules prisma cli node modules prisma engines prisma fmt debian openssl x studio
| 1
|
16,926
| 22,272,992,774
|
IssuesEvent
|
2022-06-10 14:01:38
|
hashgraph/hedera-json-rpc-relay
|
https://api.github.com/repos/hashgraph/hedera-json-rpc-relay
|
closed
|
Startup logs should inform correct setup
|
enhancement P1 process
|
### Problem
Currently the initial set of logs don't always make it clear what configurations have been set and whether correct operation of the relay is expected.
### Solution
Improve logs to clearly show
- mIrror node url
- consensus node endpoints
### Alternatives
_No response_
|
1.0
|
Startup logs should inform correct setup - ### Problem
Currently the initial set of logs don't always make it clear what configurations have been set and whether correct operation of the relay is expected.
### Solution
Improve logs to clearly show
- mIrror node url
- consensus node endpoints
### Alternatives
_No response_
|
process
|
startup logs should inform correct setup problem currently the initial set of logs don t always make it clear what configurations have been set and whether correct operation of the relay is expected solution improve logs to clearly show mirror node url consensus node endpoints alternatives no response
| 1
|
20,548
| 27,204,387,404
|
IssuesEvent
|
2023-02-20 12:01:37
|
GIScience/sketch-map-tool
|
https://api.github.com/repos/GIScience/sketch-map-tool
|
closed
|
Sketch Maps in portrait orientation are not correctly georeferenced
|
bug component:upload-processing
|
Both GeoTIFF and vector data are rotated by 90 degrees
|
1.0
|
Sketch Maps in portrait orientation are not correctly georeferenced - Both GeoTIFF and vector data are rotated by 90 degrees
|
process
|
sketch maps in portrait orientation are not correctly georeferenced both geotiff and vector data are rotated by degrees
| 1
|
12,589
| 14,991,895,956
|
IssuesEvent
|
2021-01-29 09:05:04
|
panther-labs/panther
|
https://api.github.com/repos/panther-labs/panther
|
opened
|
Define pattern/framework for Lambdas to create System Health alarms
|
p1 story team:data processing
|
### Description
Define framework for Lambdas to create System Health alarms. The same pattern/framework will be used in all parts of the system that require to set up their System Health alarms.
### Related Services
All
### Designs
Not needed
### Acceptance Criteria
- Implementation of a metrics framework that our system can use to publish metrics to CloudWatch
- The framework should make use of CloudWatch alarms
- The framework is used in one part of the system to demonstrate its usage
|
1.0
|
Define pattern/framework for Lambdas to create System Health alarms - ### Description
Define framework for Lambdas to create System Health alarms. The same pattern/framework will be used in all parts of the system that require to set up their System Health alarms.
### Related Services
All
### Designs
Not needed
### Acceptance Criteria
- Implementation of a metrics framework that our system can use to publish metrics to CloudWatch
- The framework should make use of CloudWatch alarms
- The framework is used in one part of the system to demonstrate its usage
|
process
|
define pattern framework for lambdas to create system health alarms description define framework for lambdas to create system health alarms the same pattern framework will be used in all parts of the system that require to set up their system health alarms related services all designs not needed acceptance criteria implementation of a metrics framework that our system can use to publish metrics to cloudwatch the framework should make use of cloudwatch alarms the framework is used in one part of the system to demonstrate its usage
| 1
|
18,181
| 24,233,374,486
|
IssuesEvent
|
2022-09-26 20:24:37
|
GoogleCloudPlatform/cloud-ops-sandbox
|
https://api.github.com/repos/GoogleCloudPlatform/cloud-ops-sandbox
|
closed
|
Dependency Dashboard
|
priority: p2 type: process
|
This issue lists Renovate updates and detected dependencies. Read the [Dependency Dashboard](https://docs.renovatebot.com/key-concepts/dashboard/) docs to learn more.
## Repository problems
These problems occurred while renovating this repository.
- WARN: Base branch does not exist - skipping
This repository currently has no open or pending branches.
## Detected dependencies
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
|
1.0
|
Dependency Dashboard - This issue lists Renovate updates and detected dependencies. Read the [Dependency Dashboard](https://docs.renovatebot.com/key-concepts/dashboard/) docs to learn more.
## Repository problems
These problems occurred while renovating this repository.
- WARN: Base branch does not exist - skipping
This repository currently has no open or pending branches.
## Detected dependencies
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
|
process
|
dependency dashboard this issue lists renovate updates and detected dependencies read the docs to learn more repository problems these problems occurred while renovating this repository warn base branch does not exist skipping this repository currently has no open or pending branches detected dependencies check this box to trigger a request for renovate to run again on this repository
| 1
|
534,315
| 15,614,194,124
|
IssuesEvent
|
2021-03-19 17:25:50
|
canonical-web-and-design/ubuntu.com
|
https://api.github.com/repos/canonical-web-and-design/ubuntu.com
|
closed
|
Take down /16-04/gcp page
|
Priority: High
|
Can you take this page down? It's not accurate and we still need to update and confirm the page copy with various teams.
---
*Reported from: https://ubuntu.com/16-04/gcp*
|
1.0
|
Take down /16-04/gcp page - Can you take this page down? It's not accurate and we still need to update and confirm the page copy with various teams.
---
*Reported from: https://ubuntu.com/16-04/gcp*
|
non_process
|
take down gcp page can you take this page down it s not accurate and we still need to update and confirm the page copy with various teams reported from
| 0
|
22,403
| 31,142,291,001
|
IssuesEvent
|
2023-08-16 01:44:39
|
cypress-io/cypress
|
https://api.github.com/repos/cypress-io/cypress
|
closed
|
Flaky test: AssertionError: Timed out retrying after 10000ms: Expected to find content: 'Spec not found' but never did.
|
OS: linux process: flaky test topic: flake ❄️ stage: flake stale
|
### Link to dashboard or CircleCI failure
https://app.circleci.com/pipelines/github/cypress-io/cypress/41301/workflows/ed35f5b9-63a5-409c-8893-f0cd8a5bf952/jobs/1709537
### Link to failing test in GitHub
https://github.com/cypress-io/cypress/blob/develop/packages/app/cypress/e2e/cypress-in-cypress-component.cy.ts#L81
### Analysis
<img width="1131" alt="Screen Shot 2022-08-05 at 12 43 53 PM" src="https://user-images.githubusercontent.com/26726429/183149599-341591f6-c557-4ea4-929f-b11e4e40ff43.png">
### Cypress Version
10.4.0
### Other
Search for this issue number in the codebase to find the test(s) skipped until this issue is fixed
|
1.0
|
Flaky test: AssertionError: Timed out retrying after 10000ms: Expected to find content: 'Spec not found' but never did. - ### Link to dashboard or CircleCI failure
https://app.circleci.com/pipelines/github/cypress-io/cypress/41301/workflows/ed35f5b9-63a5-409c-8893-f0cd8a5bf952/jobs/1709537
### Link to failing test in GitHub
https://github.com/cypress-io/cypress/blob/develop/packages/app/cypress/e2e/cypress-in-cypress-component.cy.ts#L81
### Analysis
<img width="1131" alt="Screen Shot 2022-08-05 at 12 43 53 PM" src="https://user-images.githubusercontent.com/26726429/183149599-341591f6-c557-4ea4-929f-b11e4e40ff43.png">
### Cypress Version
10.4.0
### Other
Search for this issue number in the codebase to find the test(s) skipped until this issue is fixed
|
process
|
flaky test assertionerror timed out retrying after expected to find content spec not found but never did link to dashboard or circleci failure link to failing test in github analysis img width alt screen shot at pm src cypress version other search for this issue number in the codebase to find the test s skipped until this issue is fixed
| 1
|
8,477
| 11,643,051,604
|
IssuesEvent
|
2020-02-29 11:05:11
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
opened
|
UCP: Migrate scalar function `SubstringIndex` from TiDB
|
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
|
## Description
Port the scalar function `SubstringIndex` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @breeswish
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
2.0
|
UCP: Migrate scalar function `SubstringIndex` from TiDB -
## Description
Port the scalar function `SubstringIndex` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @breeswish
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
process
|
ucp migrate scalar function substringindex from tidb description port the scalar function substringindex from tidb to coprocessor score mentor s breeswish recommended skills rust programming learning materials already implemented expressions ported from tidb
| 1
|
6,735
| 9,866,780,500
|
IssuesEvent
|
2019-06-21 08:31:32
|
googleapis/google-cloud-python
|
https://api.github.com/repos/googleapis/google-cloud-python
|
closed
|
Missing system tests after pubsub redesign
|
api: pubsub testing triaged for GA type: process
|
The redesign effort (#3859) left behind only very minimal system tests. The older implementation had much broader system test coverage:
- Listing topics and subscriptions in the client's project.
- Creating subscriptions with non-default settings.
- Listing subscriptions bound to a topic.
- Setting / getting IAM policy for topics and subscriptions
- Creating / seeking snapshots.
I noticed the missing tests today when I went looking for examples of updating a topic's IAM policy.
|
1.0
|
Missing system tests after pubsub redesign - The redesign effort (#3859) left behind only very minimal system tests. The older implementation had much broader system test coverage:
- Listing topics and subscriptions in the client's project.
- Creating subscriptions with non-default settings.
- Listing subscriptions bound to a topic.
- Setting / getting IAM policy for topics and subscriptions
- Creating / seeking snapshots.
I noticed the missing tests today when I went looking for examples of updating a topic's IAM policy.
|
process
|
missing system tests after pubsub redesign the redesign effort left behind only very minimal system tests the older implementation had much broader system test coverage listing topics and subscriptions in the client s project creating subscriptions with non default settings listing subscriptions bound to a topic setting getting iam policy for topics and subscriptions creating seeking snapshots i noticed the missing tests today when i went looking for examples of updating a topic s iam policy
| 1
|
99,989
| 4,075,204,750
|
IssuesEvent
|
2016-05-29 01:39:16
|
revel/revel
|
https://api.github.com/repos/revel/revel
|
closed
|
Update references to github.com/revel/config
|
priority-should topic-config type-enhancement
|
As mentioned in the documentation, update the reference in the code base as well.
- [x] First synchronize the fork from parent `github.com/robfig/config`
- [x] Update `github.com/robfig/config` => `github.com/revel/config`
|
1.0
|
Update references to github.com/revel/config - As mentioned in the documentation, update the reference in the code base as well.
- [x] First synchronize the fork from parent `github.com/robfig/config`
- [x] Update `github.com/robfig/config` => `github.com/revel/config`
|
non_process
|
update references to github com revel config as mentioned in the documentation update the reference in the code base as well first synchronize the fork from parent github com robfig config update github com robfig config github com revel config
| 0
|
6,043
| 7,469,519,908
|
IssuesEvent
|
2018-04-02 23:18:41
|
Microsoft/vscode-cpptools
|
https://api.github.com/repos/Microsoft/vscode-cpptools
|
closed
|
Please add support for intellisence for /clr
|
Feature Request Language Service fixed (release pending) quick fix
|
I have a C++ project with code that must be compiled with /clr.
Please add support for intellisence for /clr
|
1.0
|
Please add support for intellisence for /clr - I have a C++ project with code that must be compiled with /clr.
Please add support for intellisence for /clr
|
non_process
|
please add support for intellisence for clr i have a c project with code that must be compiled with clr please add support for intellisence for clr
| 0
|
15,300
| 19,325,017,210
|
IssuesEvent
|
2021-12-14 10:22:33
|
decidim/decidim
|
https://api.github.com/repos/decidim/decidim
|
opened
|
Removing a scope from selected scope picker doesn't work
|
type: bug module: participatory processes
|
**Describe the bug**
When I select a scope from the scope picker, for instance from participatory proceses, it doesn't behave correctly when refreshing
**To Reproduce**
Steps to reproduce the behavior:
1. Go to /processes
2. Click on "Select a scope"
3. Click on any scope
4. Error 1: if the page has scroll it goes to top
5. Refresh the page
6. Error 2: see that the label has lost the CSS design
7. Click on the label to remove.
8. Error 3: See that it disappears, but the filter selection doesn't change
8. Refresh the page
9. Error 4: see that the label/scope is there again
**Expected behavior**
I expect 3 things:
1. When I click in a scope in this picker and there's scroll, it shouldn't go to the top of the page (Error 1)
2. When I refresh after selecting a scope, I shouldn't lose the label's CSS (Error 2)
3. When I remove a filtered scope, it should show the change (Error 3)
4. When I refresh after removing a filtered scope, it should be removed for good (Error 4)
**Screenshots**

**Extra data (please complete the following information):**
- Device: Desktop
- Browser: Firefox
- Decidim Version: 0.26.0.dev
- Decidim installation: Codegram staging
|
1.0
|
Removing a scope from selected scope picker doesn't work - **Describe the bug**
When I select a scope from the scope picker, for instance from participatory proceses, it doesn't behave correctly when refreshing
**To Reproduce**
Steps to reproduce the behavior:
1. Go to /processes
2. Click on "Select a scope"
3. Click on any scope
4. Error 1: if the page has scroll it goes to top
5. Refresh the page
6. Error 2: see that the label has lost the CSS design
7. Click on the label to remove.
8. Error 3: See that it disappears, but the filter selection doesn't change
8. Refresh the page
9. Error 4: see that the label/scope is there again
**Expected behavior**
I expect 3 things:
1. When I click in a scope in this picker and there's scroll, it shouldn't go to the top of the page (Error 1)
2. When I refresh after selecting a scope, I shouldn't lose the label's CSS (Error 2)
3. When I remove a filtered scope, it should show the change (Error 3)
4. When I refresh after removing a filtered scope, it should be removed for good (Error 4)
**Screenshots**

**Extra data (please complete the following information):**
- Device: Desktop
- Browser: Firefox
- Decidim Version: 0.26.0.dev
- Decidim installation: Codegram staging
|
process
|
removing a scope from selected scope picker doesn t work describe the bug when i select a scope from the scope picker for instance from participatory proceses it doesn t behave correctly when refreshing to reproduce steps to reproduce the behavior go to processes click on select a scope click on any scope error if the page has scroll it goes to top refresh the page error see that the label has lost the css design click on the label to remove error see that it disappears but the filter selection doesn t change refresh the page error see that the label scope is there again expected behavior i expect things when i click in a scope in this picker and there s scroll it shouldn t go to the top of the page error when i refresh after selecting a scope i shouldn t lose the label s css error when i remove a filtered scope it should show the change error when i refresh after removing a filtered scope it should be removed for good error screenshots extra data please complete the following information device desktop browser firefox decidim version dev decidim installation codegram staging
| 1
|
11,858
| 14,665,036,366
|
IssuesEvent
|
2020-12-29 13:23:39
|
modi-w/AutoVersionsDB
|
https://api.github.com/repos/modi-w/AutoVersionsDB
|
opened
|
Handle State Data Files
|
area-Core area-Tests area-UI process-discussion type-enhancement
|
**The Problem**
The main purpose of this tool is to automate the process of setting the database in sync with the specific location (specific commit) with code on the source control. In other words: make it easy to sync the DB state with the code state no matter which branch\commit the code.
But sometimes in some projects, the database is not the only data state of the system. Sometimes we have external files that define the data state of the system.
For example: for accounting modules in the system, we have some pdf files for invoices.
**Solution**
This section should be for discussion.
One option maybe adds kind of "script folders" for:
1. DevDummyData state data files
2. Repeatable state data files
**Action Items:**
1.
2.
3.
**Updates**
1.
|
1.0
|
Handle State Data Files - **The Problem**
The main purpose of this tool is to automate the process of setting the database in sync with the specific location (specific commit) with code on the source control. In other words: make it easy to sync the DB state with the code state no matter which branch\commit the code.
But sometimes in some projects, the database is not the only data state of the system. Sometimes we have external files that define the data state of the system.
For example: for accounting modules in the system, we have some pdf files for invoices.
**Solution**
This section should be for discussion.
One option maybe adds kind of "script folders" for:
1. DevDummyData state data files
2. Repeatable state data files
**Action Items:**
1.
2.
3.
**Updates**
1.
|
process
|
handle state data files the problem the main purpose of this tool is to automate the process of setting the database in sync with the specific location specific commit with code on the source control in other words make it easy to sync the db state with the code state no matter which branch commit the code but sometimes in some projects the database is not the only data state of the system sometimes we have external files that define the data state of the system for example for accounting modules in the system we have some pdf files for invoices solution this section should be for discussion one option maybe adds kind of script folders for devdummydata state data files repeatable state data files action items updates
| 1
|
20,915
| 27,754,011,586
|
IssuesEvent
|
2023-03-15 23:54:40
|
dDevTech/tapas-top-frontend
|
https://api.github.com/repos/dDevTech/tapas-top-frontend
|
closed
|
Modificación pagina de registro 20/03/2023
|
pending in process require testing
|
Como jhipster nos ha creado las paginas por defecto, se debe modificar la de registro para que el botón Crear Cuenta sea continuar y nos lleva a la pagina register-account-info. Debemos haber guardado la información de registro con redux de forma temporal en vez de enviarla directamente a la base de datos
|
1.0
|
Modificación pagina de registro 20/03/2023 - Como jhipster nos ha creado las paginas por defecto, se debe modificar la de registro para que el botón Crear Cuenta sea continuar y nos lleva a la pagina register-account-info. Debemos haber guardado la información de registro con redux de forma temporal en vez de enviarla directamente a la base de datos
|
process
|
modificación pagina de registro como jhipster nos ha creado las paginas por defecto se debe modificar la de registro para que el botón crear cuenta sea continuar y nos lleva a la pagina register account info debemos haber guardado la información de registro con redux de forma temporal en vez de enviarla directamente a la base de datos
| 1
|
18,765
| 24,669,222,636
|
IssuesEvent
|
2022-10-18 12:40:31
|
MasterPlayer/adxl345-sv
|
https://api.github.com/repos/MasterPlayer/adxl345-sv
|
opened
|
FIFO interrupt processing support
|
enhancement hardware process software process
|
There are needed for special processing for FIFO interrupts, because sw must read series of samples.
Possible mechanism may be realized as next scheme:
|
2.0
|
FIFO interrupt processing support - There are needed for special processing for FIFO interrupts, because sw must read series of samples.
Possible mechanism may be realized as next scheme:
|
process
|
fifo interrupt processing support there are needed for special processing for fifo interrupts because sw must read series of samples possible mechanism may be realized as next scheme
| 1
|
21,077
| 28,019,962,159
|
IssuesEvent
|
2023-03-28 04:03:29
|
0xPolygonMiden/miden-vm
|
https://api.github.com/repos/0xPolygonMiden/miden-vm
|
closed
|
Replace MerkleSets in the advice provider with MerkleStore
|
processor
|
Now that we have [MerkleStore](https://github.com/0xPolygonMiden/crypto/blob/next/src/merkle/store.rs) implemented in `miden-crypto`, we should use it for instead of a map of `MerkleSet`'s in [MemAdviceProvider](https://github.com/0xPolygonMiden/miden-vm/blob/main/processor/src/advice/mem_provider.rs). We should also probable rename the fields like this:
* `tape` -> `stack`
* `values` -> `map`
* `sets` -> `store`
So, the `MemAdviceProvider` struct could look like this:
```Rust
pub struct MemAdviceProvider {
step: u32,
stack: Vec<Felt>,
map: BTreeMap<[u8; 32], Vec<Felt>>,
store: MerkleStore,
}
```
We might also want to consider renaming `tape` into `stack` in the [AdviceProvider](https://github.com/0xPolygonMiden/miden-vm/blob/main/processor/src/advice/mod.rs#L52) trait as well.
|
1.0
|
Replace MerkleSets in the advice provider with MerkleStore - Now that we have [MerkleStore](https://github.com/0xPolygonMiden/crypto/blob/next/src/merkle/store.rs) implemented in `miden-crypto`, we should use it for instead of a map of `MerkleSet`'s in [MemAdviceProvider](https://github.com/0xPolygonMiden/miden-vm/blob/main/processor/src/advice/mem_provider.rs). We should also probable rename the fields like this:
* `tape` -> `stack`
* `values` -> `map`
* `sets` -> `store`
So, the `MemAdviceProvider` struct could look like this:
```Rust
pub struct MemAdviceProvider {
step: u32,
stack: Vec<Felt>,
map: BTreeMap<[u8; 32], Vec<Felt>>,
store: MerkleStore,
}
```
We might also want to consider renaming `tape` into `stack` in the [AdviceProvider](https://github.com/0xPolygonMiden/miden-vm/blob/main/processor/src/advice/mod.rs#L52) trait as well.
|
process
|
replace merklesets in the advice provider with merklestore now that we have implemented in miden crypto we should use it for instead of a map of merkleset s in we should also probable rename the fields like this tape stack values map sets store so the memadviceprovider struct could look like this rust pub struct memadviceprovider step stack vec map btreemap store merklestore we might also want to consider renaming tape into stack in the trait as well
| 1
|
327,932
| 24,162,319,206
|
IssuesEvent
|
2022-09-22 12:39:32
|
giantswarm/roadmap
|
https://api.github.com/repos/giantswarm/roadmap
|
closed
|
Deprecate CRDs
|
topic/documentation team/rainbow topic/crd
|
With CAPI, many CRDs we currently document in [our docs](https://docs.giantswarm.io/ui-api/management-api/crd/) are no longer used.
I'd like to mark these CRDs as deprecated. This way it's easier to understand for everyone which CRDs have a future.
|
1.0
|
Deprecate CRDs - With CAPI, many CRDs we currently document in [our docs](https://docs.giantswarm.io/ui-api/management-api/crd/) are no longer used.
I'd like to mark these CRDs as deprecated. This way it's easier to understand for everyone which CRDs have a future.
|
non_process
|
deprecate crds with capi many crds we currently document in are no longer used i d like to mark these crds as deprecated this way it s easier to understand for everyone which crds have a future
| 0
|
15,308
| 19,400,850,809
|
IssuesEvent
|
2021-12-19 06:13:31
|
ethereum/EIPs
|
https://api.github.com/repos/ethereum/EIPs
|
closed
|
Add mission statement
|
type: Meta type: EIP1 (Process) stale
|
Presently the ethereum/EIPs project does not have a mission statement.
---
<strike>Recently something changed and now the majority of EIPs here have no path to become "final" standards. Pull request #1100 addresses that issue.</strike>
However, one of the EIP editors (the people with commit access here) mentioned that #1100 is not urgent. There are no remaining complaints on #1100, it has EIP editor endorsements, but it is not merged.
I reviewed the project README.md and was hoping to find something like "our goal is to discuss and pass high-quality standards reflecting established best practices in the community." So I could tell this person that #1100 is urgent (because presently, standards are prevented from passing). Alas no such line exists, in fact, there is nothing in the README.md that explains why we are contributing here.
**It is much easier to set expectations for each other in this project if we have a clearly defined goal. And we should state that goal in the README.md.**
|
1.0
|
Add mission statement - Presently the ethereum/EIPs project does not have a mission statement.
---
<strike>Recently something changed and now the majority of EIPs here have no path to become "final" standards. Pull request #1100 addresses that issue.</strike>
However, one of the EIP editors (the people with commit access here) mentioned that #1100 is not urgent. There are no remaining complaints on #1100, it has EIP editor endorsements, but it is not merged.
I reviewed the project README.md and was hoping to find something like "our goal is to discuss and pass high-quality standards reflecting established best practices in the community." So I could tell this person that #1100 is urgent (because presently, standards are prevented from passing). Alas no such line exists, in fact, there is nothing in the README.md that explains why we are contributing here.
**It is much easier to set expectations for each other in this project if we have a clearly defined goal. And we should state that goal in the README.md.**
|
process
|
add mission statement presently the ethereum eips project does not have a mission statement recently something changed and now the majority of eips here have no path to become final standards pull request addresses that issue however one of the eip editors the people with commit access here mentioned that is not urgent there are no remaining complaints on it has eip editor endorsements but it is not merged i reviewed the project readme md and was hoping to find something like our goal is to discuss and pass high quality standards reflecting established best practices in the community so i could tell this person that is urgent because presently standards are prevented from passing alas no such line exists in fact there is nothing in the readme md that explains why we are contributing here it is much easier to set expectations for each other in this project if we have a clearly defined goal and we should state that goal in the readme md
| 1
|
11,617
| 14,480,903,733
|
IssuesEvent
|
2020-12-10 11:50:57
|
Arch666Angel/mods
|
https://api.github.com/repos/Arch666Angel/mods
|
closed
|
[BUG] Agriculture Modules can be put in labs
|
Angels Bio Processing Impact: Bug
|
Agriculture Modules can be put in labs.
Tested with just Angel's Mods (no overhauls) as well as with Bob's Technology.
Issue applies to all labs. With bobs: lab, lab 2, alien lab, modules lab.
|
1.0
|
[BUG] Agriculture Modules can be put in labs - Agriculture Modules can be put in labs.
Tested with just Angel's Mods (no overhauls) as well as with Bob's Technology.
Issue applies to all labs. With bobs: lab, lab 2, alien lab, modules lab.
|
process
|
agriculture modules can be put in labs agriculture modules can be put in labs tested with just angel s mods no overhauls as well as with bob s technology issue applies to all labs with bobs lab lab alien lab modules lab
| 1
|
342,565
| 30,627,464,733
|
IssuesEvent
|
2023-07-24 12:31:58
|
unifyai/ivy
|
https://api.github.com/repos/unifyai/ivy
|
opened
|
Fix linalg.test_tensorflow_matrix_transpose
|
TensorFlow Frontend Sub Task Failing Test
|
| | |
|---|---|
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5634736017/job/15264943658"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5634736017/job/15264943658"><img src=https://img.shields.io/badge/-success-success></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5634736017/job/15264943658"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5634736017/job/15264943658"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5634736017/job/15264943658"><img src=https://img.shields.io/badge/-failure-red></a>
|
1.0
|
Fix linalg.test_tensorflow_matrix_transpose - | | |
|---|---|
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5634736017/job/15264943658"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5634736017/job/15264943658"><img src=https://img.shields.io/badge/-success-success></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5634736017/job/15264943658"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5634736017/job/15264943658"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5634736017/job/15264943658"><img src=https://img.shields.io/badge/-failure-red></a>
|
non_process
|
fix linalg test tensorflow matrix transpose jax a href src numpy a href src tensorflow a href src torch a href src paddle a href src
| 0
|
21,934
| 30,446,677,667
|
IssuesEvent
|
2023-07-15 19:08:58
|
h4sh5/pypi-auto-scanner
|
https://api.github.com/repos/h4sh5/pypi-auto-scanner
|
opened
|
pyutils 0.0.1b5 has 2 GuardDog issues
|
guarddog typosquatting silent-process-execution
|
https://pypi.org/project/pyutils
https://inspector.pypi.io/project/pyutils
```{
"dependency": "pyutils",
"version": "0.0.1b5",
"result": {
"issues": 2,
"errors": {},
"results": {
"typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: python-utils, pytils",
"silent-process-execution": [
{
"location": "pyutils-0.0.1b5/src/pyutils/exec_utils.py:204",
"code": " subproc = subprocess.Popen(\n args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n )",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmpx5rt84g_/pyutils"
}
}```
|
1.0
|
pyutils 0.0.1b5 has 2 GuardDog issues - https://pypi.org/project/pyutils
https://inspector.pypi.io/project/pyutils
```{
"dependency": "pyutils",
"version": "0.0.1b5",
"result": {
"issues": 2,
"errors": {},
"results": {
"typosquatting": "This package closely ressembles the following package names, and might be a typosquatting attempt: python-utils, pytils",
"silent-process-execution": [
{
"location": "pyutils-0.0.1b5/src/pyutils/exec_utils.py:204",
"code": " subproc = subprocess.Popen(\n args,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n )",
"message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null"
}
]
},
"path": "/tmp/tmpx5rt84g_/pyutils"
}
}```
|
process
|
pyutils has guarddog issues dependency pyutils version result issues errors results typosquatting this package closely ressembles the following package names and might be a typosquatting attempt python utils pytils silent process execution location pyutils src pyutils exec utils py code subproc subprocess popen n args n stdin subprocess devnull n stdout subprocess devnull n stderr subprocess devnull n message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp pyutils
| 1
|
3,901
| 6,822,593,502
|
IssuesEvent
|
2017-11-07 20:38:05
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
Improve ChildProcess::killed property behaviour. Or update documentation.
|
child_process
|
<!--
Thank you for reporting an issue.
This issue tracker is for bugs and issues found within Node.js core.
If you require more general support please file an issue on our help
repo. https://github.com/nodejs/help
Please fill in as much of the template below as you're able.
Subsystem: doc
If possible, please provide code that demonstrates the problem, keeping it as
simple and free of external dependencies as you are able.
-->
* **Subsystem**: doc and child_process
<!-- Enter your issue details below this comment. -->
According to documentation (https://nodejs.org/dist/latest-v9.x/docs/api/child_process.html#child_process_subprocess_killed) the `killed` property has invalid semantic in comparison to ChildProcess::kill method behaviour (internal/child_process.js)
Concretely, the problematic code is
```
var err = this._handle.kill(signal);
if (err === 0) {
/* Success. */
this.killed = true;
return true;
}
```
After the `this._handle.kill(signal);` method call we have no idea whether the 3-d party process was killed, or it just received a signal, processed it and continues working (it maybe be possible even with such 'kill' signals like SIGINT and SIGTERM).
So I propose to discuss what can be done in this direction to improve expected behaviour.
The easiest solution (and maybe the right one) is to move `this.killed = true;` line into `_handle.onexit` function and fix to the following code:
```
if (signalCode) {
this.signalCode = signalCode;
if (this._killWasCalled) { (1)
this.killed = true;
}
} else {
this.exitCode = exitCode;
}
```
The if block (1) is needed cuz now documentation says '...indicates whether the child process was successfully terminated using `subprocess.kill()`'. We can eliminate this block, but we will need to update documentation also.
Strictly speaking, the `killed` field isn't very helpful in terms of identifying when process will exit.
We usually use `exit`event for that. But, nevertheless, it's strange to see childProcess.killed === true, when in reality process works fine.
|
1.0
|
Improve ChildProcess::killed property behaviour. Or update documentation. - <!--
Thank you for reporting an issue.
This issue tracker is for bugs and issues found within Node.js core.
If you require more general support please file an issue on our help
repo. https://github.com/nodejs/help
Please fill in as much of the template below as you're able.
Subsystem: doc
If possible, please provide code that demonstrates the problem, keeping it as
simple and free of external dependencies as you are able.
-->
* **Subsystem**: doc and child_process
<!-- Enter your issue details below this comment. -->
According to documentation (https://nodejs.org/dist/latest-v9.x/docs/api/child_process.html#child_process_subprocess_killed) the `killed` property has invalid semantic in comparison to ChildProcess::kill method behaviour (internal/child_process.js)
Concretely, the problematic code is
```
var err = this._handle.kill(signal);
if (err === 0) {
/* Success. */
this.killed = true;
return true;
}
```
After the `this._handle.kill(signal);` method call we have no idea whether the 3-d party process was killed, or it just received a signal, processed it and continues working (it maybe be possible even with such 'kill' signals like SIGINT and SIGTERM).
So I propose to discuss what can be done in this direction to improve expected behaviour.
The easiest solution (and maybe the right one) is to move `this.killed = true;` line into `_handle.onexit` function and fix to the following code:
```
if (signalCode) {
this.signalCode = signalCode;
if (this._killWasCalled) { (1)
this.killed = true;
}
} else {
this.exitCode = exitCode;
}
```
The if block (1) is needed cuz now documentation says '...indicates whether the child process was successfully terminated using `subprocess.kill()`'. We can eliminate this block, but we will need to update documentation also.
Strictly speaking, the `killed` field isn't very helpful in terms of identifying when process will exit.
We usually use `exit`event for that. But, nevertheless, it's strange to see childProcess.killed === true, when in reality process works fine.
|
process
|
improve childprocess killed property behaviour or update documentation thank you for reporting an issue this issue tracker is for bugs and issues found within node js core if you require more general support please file an issue on our help repo please fill in as much of the template below as you re able subsystem doc if possible please provide code that demonstrates the problem keeping it as simple and free of external dependencies as you are able subsystem doc and child process according to documentation the killed property has invalid semantic in comparison to childprocess kill method behaviour internal child process js concretely the problematic code is var err this handle kill signal if err success this killed true return true after the this handle kill signal method call we have no idea whether the d party process was killed or it just received a signal processed it and continues working it maybe be possible even with such kill signals like sigint and sigterm so i propose to discuss what can be done in this direction to improve expected behaviour the easiest solution and maybe the right one is to move this killed true line into handle onexit function and fix to the following code if signalcode this signalcode signalcode if this killwascalled this killed true else this exitcode exitcode the if block is needed cuz now documentation says indicates whether the child process was successfully terminated using subprocess kill we can eliminate this block but we will need to update documentation also strictly speaking the killed field isn t very helpful in terms of identifying when process will exit we usually use exit event for that but nevertheless it s strange to see childprocess killed true when in reality process works fine
| 1
|
245,003
| 26,498,586,526
|
IssuesEvent
|
2023-01-18 08:30:05
|
dedis/d-voting
|
https://api.github.com/repos/dedis/d-voting
|
closed
|
THREAT - A user who is not an admin or operator cannot vote.
|
security issue web backend
|
## Scenario
Every action from a user who is not an admin or operator will become unauthorized, including casting a vote. An authenticated user should able to cast a vote even if they are not an operator or admin
## Source
web/backend/src/Server.ts
```js
// Secure /api/evoting to admins and operators
app.use('/api/evoting/*', (req, res, next) => {
if (!isAuthorized(req.session.userid, SUBJECT_ELECTION, ACTION_CREATE)) {
res.status(400).send('Unauthorized - only admins and operators allowed');
return;
}
next();
});
```
## Breaking Property
Availability, Authorization
## Risk
CVSS Score: [5.7/10](https://www.first.org/cvss/calculator/3.1#CVSS:3.1/AV:N/AC:L/PR:L/UI:R/S:U/C:N/I:N/A:H)
## Mitigation
Backend should let an authenticated user have the ability to vote.
|
True
|
THREAT - A user who is not an admin or operator cannot vote. - ## Scenario
Every action from a user who is not an admin or operator will become unauthorized, including casting a vote. An authenticated user should able to cast a vote even if they are not an operator or admin
## Source
web/backend/src/Server.ts
```js
// Secure /api/evoting to admins and operators
app.use('/api/evoting/*', (req, res, next) => {
if (!isAuthorized(req.session.userid, SUBJECT_ELECTION, ACTION_CREATE)) {
res.status(400).send('Unauthorized - only admins and operators allowed');
return;
}
next();
});
```
## Breaking Property
Availability, Authorization
## Risk
CVSS Score: [5.7/10](https://www.first.org/cvss/calculator/3.1#CVSS:3.1/AV:N/AC:L/PR:L/UI:R/S:U/C:N/I:N/A:H)
## Mitigation
Backend should let an authenticated user have the ability to vote.
|
non_process
|
threat a user who is not an admin or operator cannot vote scenario every action from a user who is not an admin or operator will become unauthorized including casting a vote an authenticated user should able to cast a vote even if they are not an operator or admin source web backend src server ts js secure api evoting to admins and operators app use api evoting req res next if isauthorized req session userid subject election action create res status send unauthorized only admins and operators allowed return next breaking property availability authorization risk cvss score mitigation backend should let an authenticated user have the ability to vote
| 0
|
11,251
| 14,018,637,091
|
IssuesEvent
|
2020-10-29 17:05:15
|
fluent/fluent-bit
|
https://api.github.com/repos/fluent/fluent-bit
|
closed
|
parser: support subsecond resolution with colon (%s:%L)
|
work-in-process
|
## Bug Report
**Describe the bug**
I have a 3rd party application (Forgerock OpenAM) that writes timestamps as follows in some log files:
```
amAuthInternal:05/29/2020 11:58:21:127 AM UTC: Thread[localhost-startStop-1,5,main]: TransactionId[3da35379-317e-41fe-b9f2-0cae565a8480-0] AuthContext::getLoginStatus()
```
Note that it uses a colon to separate subseconds, hence of the format "%S:%L". As per https://github.com/fluent/fluent-bit/issues/703, only "%S.%L" (dot) and "%S,%L" (comma) are currently supported by fluent-bit.
**To Reproduce**
Consider log entries as follows:
```
amAuthInternal:05/29/2020 11:58:21:127 AM UTC: Thread[localhost-startStop-1,5,main]: TransactionId[3da35379-317e-41fe-b9f2-0cae565a8480-0] AuthContext::getLoginStatus()
amAuthInternal:05/29/2020 11:58:21:128 AM UTC: Thread[localhost-startStop-1,5,main]: TransactionId[3da35379-317e-41fe-b9f2-0cae565a8480-0] AuthContext::getSubject()
amAuthInternal:05/29/2020 11:58:21:128 AM UTC: Thread[localhost-startStop-1,5,main]: TransactionId[3da35379-317e-41fe-b9f2-0cae565a8480-0] AuthContext::getAuthPrincipal(): [AuthPrincipal: cn=dsameuser,ou=DSAME Users,ou=am-config]
```
Parser config:
```
[PARSER]
Name am-debug
Format regex
Regex ^(?<module>[A-Za-z0-9]*):(?<timestamp>\d{2}/\d{2}/\d{4} \d{2}:\d{2}:\d{2}:\d{3} [^\ ]* [^:]*): Thread\[(?<thread>[^\]]*)\]: TransactionId\[(?<transactionId>[^\]]*)\] (?<message>.*)$
Time_Key timestamp
Time_Format %m/%d/%Y %I:%M:%S:%L %p %Z
Time_Keep On
```
The result:
```
[2020/05/29 12:10:37] [ warn] [parser:am-debug] Invalid time format %m/%d/%Y %I:%M:%S:%L %p %Z for '05/29/2020 12:10:37:601 PM UTC'.
```
**Expected behavior**
It should be possible to have any separator between %S and %L. This should be defined via Time_Format.
**Your Environment**
* Version used: 1.4.5
* Filters and plugins: tail
|
1.0
|
parser: support subsecond resolution with colon (%s:%L) - ## Bug Report
**Describe the bug**
I have a 3rd party application (Forgerock OpenAM) that writes timestamps as follows in some log files:
```
amAuthInternal:05/29/2020 11:58:21:127 AM UTC: Thread[localhost-startStop-1,5,main]: TransactionId[3da35379-317e-41fe-b9f2-0cae565a8480-0] AuthContext::getLoginStatus()
```
Note that it uses a colon to separate subseconds, hence of the format "%S:%L". As per https://github.com/fluent/fluent-bit/issues/703, only "%S.%L" (dot) and "%S,%L" (comma) are currently supported by fluent-bit.
**To Reproduce**
Consider log entries as follows:
```
amAuthInternal:05/29/2020 11:58:21:127 AM UTC: Thread[localhost-startStop-1,5,main]: TransactionId[3da35379-317e-41fe-b9f2-0cae565a8480-0] AuthContext::getLoginStatus()
amAuthInternal:05/29/2020 11:58:21:128 AM UTC: Thread[localhost-startStop-1,5,main]: TransactionId[3da35379-317e-41fe-b9f2-0cae565a8480-0] AuthContext::getSubject()
amAuthInternal:05/29/2020 11:58:21:128 AM UTC: Thread[localhost-startStop-1,5,main]: TransactionId[3da35379-317e-41fe-b9f2-0cae565a8480-0] AuthContext::getAuthPrincipal(): [AuthPrincipal: cn=dsameuser,ou=DSAME Users,ou=am-config]
```
Parser config:
```
[PARSER]
Name am-debug
Format regex
Regex ^(?<module>[A-Za-z0-9]*):(?<timestamp>\d{2}/\d{2}/\d{4} \d{2}:\d{2}:\d{2}:\d{3} [^\ ]* [^:]*): Thread\[(?<thread>[^\]]*)\]: TransactionId\[(?<transactionId>[^\]]*)\] (?<message>.*)$
Time_Key timestamp
Time_Format %m/%d/%Y %I:%M:%S:%L %p %Z
Time_Keep On
```
The result:
```
[2020/05/29 12:10:37] [ warn] [parser:am-debug] Invalid time format %m/%d/%Y %I:%M:%S:%L %p %Z for '05/29/2020 12:10:37:601 PM UTC'.
```
**Expected behavior**
It should be possible to have any separator between %S and %L. This should be defined via Time_Format.
**Your Environment**
* Version used: 1.4.5
* Filters and plugins: tail
|
process
|
parser support subsecond resolution with colon s l bug report describe the bug i have a party application forgerock openam that writes timestamps as follows in some log files amauthinternal am utc thread transactionid authcontext getloginstatus note that it uses a colon to separate subseconds hence of the format s l as per only s l dot and s l comma are currently supported by fluent bit to reproduce consider log entries as follows amauthinternal am utc thread transactionid authcontext getloginstatus amauthinternal am utc thread transactionid authcontext getsubject amauthinternal am utc thread transactionid authcontext getauthprincipal parser config name am debug format regex regex d d d d d d d thread transactionid time key timestamp time format m d y i m s l p z time keep on the result invalid time format m d y i m s l p z for pm utc expected behavior it should be possible to have any separator between s and l this should be defined via time format your environment version used filters and plugins tail
| 1
|
1,304
| 3,857,339,171
|
IssuesEvent
|
2016-04-07 05:18:49
|
PlagueHO/LabBuilder
|
https://api.github.com/repos/PlagueHO/LabBuilder
|
closed
|
Change Nano Server Package Property to expect actual package names
|
enhancement In Process
|
Currently the packages specified in the Packages property in a Nano Server VM configuration are looked up in an internal array and mapped to the actual filenames of the packages that are found on the ISO. This is not generic and will require updating every time new packages are released.
This change should ensure that the actual filename of the package should be able to be specified instead.
|
1.0
|
Change Nano Server Package Property to expect actual package names - Currently the packages specified in the Packages property in a Nano Server VM configuration are looked up in an internal array and mapped to the actual filenames of the packages that are found on the ISO. This is not generic and will require updating every time new packages are released.
This change should ensure that the actual filename of the package should be able to be specified instead.
|
process
|
change nano server package property to expect actual package names currently the packages specified in the packages property in a nano server vm configuration are looked up in an internal array and mapped to the actual filenames of the packages that are found on the iso this is not generic and will require updating every time new packages are released this change should ensure that the actual filename of the package should be able to be specified instead
| 1
|
25,996
| 12,339,624,335
|
IssuesEvent
|
2020-05-14 18:27:05
|
elastic/kibana
|
https://api.github.com/repos/elastic/kibana
|
opened
|
[Alerting] [Discuss] Modifying alert params within the executor as a migration tool
|
Feature:Alerting Team:Alerting Services discuss
|
Alerts are in beta right now, so we might be pushing a large amount of breaking changes to them. Required alert params might change between minor versions, and so alerts created in a previous version would immediately break.
Can we fix this by including a migration assistant service in alert executors? For example:
- Metric alert types have an interval (`timeSize` and `timeUnit`) inside each alert `criteria` (an array of alert conditions), but these are supposed to be the same for each of them. It's redundant and error-prone to repeat the same value inside each array member.
```js
params: {
criteria: [{
timeSize: 's' | 'm' | 'h' | 'd',
timeUnit: number,
...rest
},
...additionalConditions
]
}
```
- In the next minor version, we might move `timeSize` and `timeUnit` outside of the `criteria` array and into the top-level alert `params`.
```js
params: {
criteria: [...arrayOfConditions],
timeSize,
timeUnit
}
```
- When the alert executor runs, it checks to see if `timeSize` and `timeUnit` are inside of the `criteria`. If they are, it will move them into the top level of params and **update the SavedObject of the alert** to reflect this change. This code can then be deprecated in the next minor version, assuming that all alerts probably migrated before then.
Is something like this already possible with the available `savedObjects` service in alerts, or would a new one be needed? Is this even a good idea? I'd like some opinions.
|
1.0
|
[Alerting] [Discuss] Modifying alert params within the executor as a migration tool - Alerts are in beta right now, so we might be pushing a large amount of breaking changes to them. Required alert params might change between minor versions, and so alerts created in a previous version would immediately break.
Can we fix this by including a migration assistant service in alert executors? For example:
- Metric alert types have an interval (`timeSize` and `timeUnit`) inside each alert `criteria` (an array of alert conditions), but these are supposed to be the same for each of them. It's redundant and error-prone to repeat the same value inside each array member.
```js
params: {
criteria: [{
timeSize: 's' | 'm' | 'h' | 'd',
timeUnit: number,
...rest
},
...additionalConditions
]
}
```
- In the next minor version, we might move `timeSize` and `timeUnit` outside of the `criteria` array and into the top-level alert `params`.
```js
params: {
criteria: [...arrayOfConditions],
timeSize,
timeUnit
}
```
- When the alert executor runs, it checks to see if `timeSize` and `timeUnit` are inside of the `criteria`. If they are, it will move them into the top level of params and **update the SavedObject of the alert** to reflect this change. This code can then be deprecated in the next minor version, assuming that all alerts probably migrated before then.
Is something like this already possible with the available `savedObjects` service in alerts, or would a new one be needed? Is this even a good idea? I'd like some opinions.
|
non_process
|
modifying alert params within the executor as a migration tool alerts are in beta right now so we might be pushing a large amount of breaking changes to them required alert params might change between minor versions and so alerts created in a previous version would immediately break can we fix this by including a migration assistant service in alert executors for example metric alert types have an interval timesize and timeunit inside each alert criteria an array of alert conditions but these are supposed to be the same for each of them it s redundant and error prone to repeat the same value inside each array member js params criteria timesize s m h d timeunit number rest additionalconditions in the next minor version we might move timesize and timeunit outside of the criteria array and into the top level alert params js params criteria timesize timeunit when the alert executor runs it checks to see if timesize and timeunit are inside of the criteria if they are it will move them into the top level of params and update the savedobject of the alert to reflect this change this code can then be deprecated in the next minor version assuming that all alerts probably migrated before then is something like this already possible with the available savedobjects service in alerts or would a new one be needed is this even a good idea i d like some opinions
| 0
|
11,120
| 13,957,685,093
|
IssuesEvent
|
2020-10-24 08:08:31
|
alexanderkotsev/geoportal
|
https://api.github.com/repos/alexanderkotsev/geoportal
|
opened
|
PL: Unsuccessful harvesting
|
Geoportal Harvesting process PL - Poland
|
Hi,
I did 2 harvesting tests, I received an e-mail confirming the start, the next day wanting to check the results and publish them, I get a new harvesting option without previous results and the possibility of publication.
Regards
Piotr
|
1.0
|
PL: Unsuccessful harvesting - Hi,
I did 2 harvesting tests, I received an e-mail confirming the start, the next day wanting to check the results and publish them, I get a new harvesting option without previous results and the possibility of publication.
Regards
Piotr
|
process
|
pl unsuccessful harvesting hi i did harvesting tests i received an e mail confirming the start the next day wanting to check the results and publish them i get a new harvesting option without previous results and the possibility of publication regards piotr
| 1
|
11,682
| 8,467,515,872
|
IssuesEvent
|
2018-10-23 17:10:54
|
AOSC-Dev/aosc-os-abbs
|
https://api.github.com/repos/AOSC-Dev/aosc-os-abbs
|
closed
|
requests: CVE-2018-18074
|
security to-stable
|
<!-- Please remove items do not apply. -->
**CVE IDs:** CVE-2018-18074
**Other security advisory IDs:** USN-3790-1
**Descriptions:**
Requests could be made to expose sensitive information if it
received a specially crafted HTTP header.
**Patches:** https://github.com/requests/requests/commit/c45d7c49ea75133e52ab22a8e9e13173938e36ff
**PoC(s):** https://github.com/requests/requests/issues/4716
**Architectural progress:**
<!-- Please remove any architecture to which the security vulnerabilities do not apply. -->
- [x] Data/architecture-independent (`noarch`)
|
True
|
requests: CVE-2018-18074 - <!-- Please remove items do not apply. -->
**CVE IDs:** CVE-2018-18074
**Other security advisory IDs:** USN-3790-1
**Descriptions:**
Requests could be made to expose sensitive information if it
received a specially crafted HTTP header.
**Patches:** https://github.com/requests/requests/commit/c45d7c49ea75133e52ab22a8e9e13173938e36ff
**PoC(s):** https://github.com/requests/requests/issues/4716
**Architectural progress:**
<!-- Please remove any architecture to which the security vulnerabilities do not apply. -->
- [x] Data/architecture-independent (`noarch`)
|
non_process
|
requests cve cve ids cve other security advisory ids usn descriptions requests could be made to expose sensitive information if it received a specially crafted http header patches poc s architectural progress data architecture independent noarch
| 0
|
9,947
| 12,976,360,272
|
IssuesEvent
|
2020-07-21 18:38:17
|
googleapis/java-game-servers
|
https://api.github.com/repos/googleapis/java-game-servers
|
closed
|
Promote to Beta
|
api: gameservices status: blocked type: process
|
Package name: **google-cloud-gameservices**
Current release: **alpha**
Proposed release: **beta**
## Instructions
Check the lists below, adding tests / documentation as required. Once all the "required" boxes are ticked, please create a release and close this issue.
## Required
- [ ] Server API is beta or GA
- [ ] Service API is public
- [ ] Client surface is mostly stable (no known issues that could significantly change the surface)
- [ ] All manual types and methods have comment documentation
- [ ] Package name is idiomatic for the platform
- [ ] At least one integration/smoke test is defined and passing
- [ ] Central GitHub README lists and points to the per-API README
- [ ] Per-API README links to product page on cloud.google.com
- [ ] Manual code has been reviewed for API stability by repo owner
## Optional
- [ ] Most common / important scenarios have descriptive samples
- [ ] Public manual methods have at least one usage sample each (excluding overloads)
- [ ] Per-API README includes a full description of the API
- [ ] Per-API README contains at least one “getting started” sample using the most common API scenario
- [ ] Manual code has been reviewed by API producer
- [ ] Manual code has been reviewed by a DPE responsible for samples
- [ ] 'Client LIbraries' page is added to the product documentation in 'APIs & Reference' section of the product's documentation on Cloud Site
|
1.0
|
Promote to Beta - Package name: **google-cloud-gameservices**
Current release: **alpha**
Proposed release: **beta**
## Instructions
Check the lists below, adding tests / documentation as required. Once all the "required" boxes are ticked, please create a release and close this issue.
## Required
- [ ] Server API is beta or GA
- [ ] Service API is public
- [ ] Client surface is mostly stable (no known issues that could significantly change the surface)
- [ ] All manual types and methods have comment documentation
- [ ] Package name is idiomatic for the platform
- [ ] At least one integration/smoke test is defined and passing
- [ ] Central GitHub README lists and points to the per-API README
- [ ] Per-API README links to product page on cloud.google.com
- [ ] Manual code has been reviewed for API stability by repo owner
## Optional
- [ ] Most common / important scenarios have descriptive samples
- [ ] Public manual methods have at least one usage sample each (excluding overloads)
- [ ] Per-API README includes a full description of the API
- [ ] Per-API README contains at least one “getting started” sample using the most common API scenario
- [ ] Manual code has been reviewed by API producer
- [ ] Manual code has been reviewed by a DPE responsible for samples
- [ ] 'Client LIbraries' page is added to the product documentation in 'APIs & Reference' section of the product's documentation on Cloud Site
|
process
|
promote to beta package name google cloud gameservices current release alpha proposed release beta instructions check the lists below adding tests documentation as required once all the required boxes are ticked please create a release and close this issue required server api is beta or ga service api is public client surface is mostly stable no known issues that could significantly change the surface all manual types and methods have comment documentation package name is idiomatic for the platform at least one integration smoke test is defined and passing central github readme lists and points to the per api readme per api readme links to product page on cloud google com manual code has been reviewed for api stability by repo owner optional most common important scenarios have descriptive samples public manual methods have at least one usage sample each excluding overloads per api readme includes a full description of the api per api readme contains at least one “getting started” sample using the most common api scenario manual code has been reviewed by api producer manual code has been reviewed by a dpe responsible for samples client libraries page is added to the product documentation in apis reference section of the product s documentation on cloud site
| 1
|
14,502
| 17,604,346,935
|
IssuesEvent
|
2021-08-17 15:16:50
|
qgis/QGIS-Documentation
|
https://api.github.com/repos/qgis/QGIS-Documentation
|
closed
|
[feature] add Create normal raster algorithm
|
Automatic new feature Processing Alg 3.14
|
Original commit: https://github.com/qgis/QGIS/commit/d6014d5dfe7d3d2754f1853c313ee16f4ee71721 by nyalldawson
Unfortunately this naughty coder did not write a description... :-(
|
1.0
|
[feature] add Create normal raster algorithm - Original commit: https://github.com/qgis/QGIS/commit/d6014d5dfe7d3d2754f1853c313ee16f4ee71721 by nyalldawson
Unfortunately this naughty coder did not write a description... :-(
|
process
|
add create normal raster algorithm original commit by nyalldawson unfortunately this naughty coder did not write a description
| 1
|
20,821
| 27,579,244,583
|
IssuesEvent
|
2023-03-08 15:08:16
|
ukri-excalibur/excalibur-tests
|
https://api.github.com/repos/ukri-excalibur/excalibur-tests
|
opened
|
Scripts for creating plots/tables of machine benchmarking
|
UCL postprocessing
|
Or "use case 3" in https://github.com/ukri-excalibur/excalibur-tests/issues/70#issue-1522882139 , duplicated here:
> Several benchmark apps, run on the same machine. Again, technically the same as the first case, but might benefit from simplified script.
|
1.0
|
Scripts for creating plots/tables of machine benchmarking - Or "use case 3" in https://github.com/ukri-excalibur/excalibur-tests/issues/70#issue-1522882139 , duplicated here:
> Several benchmark apps, run on the same machine. Again, technically the same as the first case, but might benefit from simplified script.
|
process
|
scripts for creating plots tables of machine benchmarking or use case in duplicated here several benchmark apps run on the same machine again technically the same as the first case but might benefit from simplified script
| 1
|
12,762
| 15,116,341,615
|
IssuesEvent
|
2021-02-09 06:33:41
|
yuta252/startlens_learning
|
https://api.github.com/repos/yuta252/startlens_learning
|
closed
|
S3 storageから学習用画像ファイルの取得
|
dev process
|
## 概要
DeepLearningモデルで学習するためには、StartlensアプリでS3に格納した画像ファイルを取得してモデルの入力層に渡す必要がある。そこでS3から画像を取得しハンドリングするためのclassを作成した。
## 変更点
- フォルダ構成と初期設定
* settings.ini, settings.pyによる設定情報の管理
* Dockerfile, docker-compose.ymlによるインフラ環境の構築
* 環境変数.envファイルの設置
* constants.pyによる定数情報の管理
- fetch/resource.pyによるS3リソースのハンドリング
## 備考
- S3からの画像取得操作はboto3のboto3.resource(高レベルAPI)もしくはboto3.client(低レベルAPI)を利用する
- S3から取得した学習用画像ファイルは、大容量になることが予想されるため、一時ファイルとして保存するのではなくByteIOを活用し、オンメモリ上で画像の加工編集を行う。
## 課題
- オンメモリ上での画像処理と学習モデルへのインプットを行うが、学習を進めていくとメモリが逼迫する可能性がある。
- そこで、明示的にガベージコレクションを利用するなどしてメモリ管理の観点からの改良を図る。
## 参照
- [boto3公式ドキュメント](https://boto3.amazonaws.com/v1/documentation/api/latest/guide/s3-example-download-file.html)
- [boto3の認証情報検索順序](https://qiita.com/tsukamoto/items/00ec8ef7e9a4ce4fb0e9)
- [boto3 で S3 から指定した prefix のオブジェクトをダウンロードする](https://rriifftt.hatenablog.com/entry/2018/06/04/144906)
- [Python の boto3 で S3 とダウンロード/アップロードする](https://sig9.hatenablog.com/entry/2020/02/02/000000)
|
1.0
|
S3 storageから学習用画像ファイルの取得 - ## 概要
DeepLearningモデルで学習するためには、StartlensアプリでS3に格納した画像ファイルを取得してモデルの入力層に渡す必要がある。そこでS3から画像を取得しハンドリングするためのclassを作成した。
## 変更点
- フォルダ構成と初期設定
* settings.ini, settings.pyによる設定情報の管理
* Dockerfile, docker-compose.ymlによるインフラ環境の構築
* 環境変数.envファイルの設置
* constants.pyによる定数情報の管理
- fetch/resource.pyによるS3リソースのハンドリング
## 備考
- S3からの画像取得操作はboto3のboto3.resource(高レベルAPI)もしくはboto3.client(低レベルAPI)を利用する
- S3から取得した学習用画像ファイルは、大容量になることが予想されるため、一時ファイルとして保存するのではなくByteIOを活用し、オンメモリ上で画像の加工編集を行う。
## 課題
- オンメモリ上での画像処理と学習モデルへのインプットを行うが、学習を進めていくとメモリが逼迫する可能性がある。
- そこで、明示的にガベージコレクションを利用するなどしてメモリ管理の観点からの改良を図る。
## 参照
- [boto3公式ドキュメント](https://boto3.amazonaws.com/v1/documentation/api/latest/guide/s3-example-download-file.html)
- [boto3の認証情報検索順序](https://qiita.com/tsukamoto/items/00ec8ef7e9a4ce4fb0e9)
- [boto3 で S3 から指定した prefix のオブジェクトをダウンロードする](https://rriifftt.hatenablog.com/entry/2018/06/04/144906)
- [Python の boto3 で S3 とダウンロード/アップロードする](https://sig9.hatenablog.com/entry/2020/02/02/000000)
|
process
|
storageから学習用画像ファイルの取得 概要 deeplearningモデルで学習するためには、 。 。 変更点 フォルダ構成と初期設定 settings ini settings pyによる設定情報の管理 dockerfile docker compose ymlによるインフラ環境の構築 環境変数 envファイルの設置 constants pyによる定数情報の管理 fetch resource 備考 resource 高レベルapi) client(低レベルapi)を利用する 、大容量になることが予想されるため、一時ファイルとして保存するのではなくbyteioを活用し、オンメモリ上で画像の加工編集を行う。 課題 オンメモリ上での画像処理と学習モデルへのインプットを行うが、学習を進めていくとメモリが逼迫する可能性がある。 そこで、明示的にガベージコレクションを利用するなどしてメモリ管理の観点からの改良を図る。 参照
| 1
|
2,991
| 5,968,535,226
|
IssuesEvent
|
2017-05-30 18:19:08
|
IFPB-2017-1/seminario
|
https://api.github.com/repos/IFPB-2017-1/seminario
|
closed
|
Criar TAP - Termo de abertura de projeto
|
Concluido Em processo
|
Seguir o modelo que está no Google Drive da professora para criar o documento inicial do projeto. Ele conterá as atribuições de cada membro do grupo e será o documento de visão com a descrição do projeto.
|
1.0
|
Criar TAP - Termo de abertura de projeto - Seguir o modelo que está no Google Drive da professora para criar o documento inicial do projeto. Ele conterá as atribuições de cada membro do grupo e será o documento de visão com a descrição do projeto.
|
process
|
criar tap termo de abertura de projeto seguir o modelo que está no google drive da professora para criar o documento inicial do projeto ele conterá as atribuições de cada membro do grupo e será o documento de visão com a descrição do projeto
| 1
|
19,719
| 10,419,850,159
|
IssuesEvent
|
2019-09-15 19:36:15
|
andOTP/andOTP
|
https://api.github.com/repos/andOTP/andOTP
|
closed
|
Bad crypto implementation?
|
question security
|
I just saw the comments in this [Reddit thread](https://old.reddit.com/r/androidapps/comments/b45zrj/dev_aegis_authenticator_secure_two_factor/ejvioko/?context=2).
1. Have the bad cryptography designs been addressed in the latest version?
2. If they havnt, have they been at least documented somewhere?
|
True
|
Bad crypto implementation? - I just saw the comments in this [Reddit thread](https://old.reddit.com/r/androidapps/comments/b45zrj/dev_aegis_authenticator_secure_two_factor/ejvioko/?context=2).
1. Have the bad cryptography designs been addressed in the latest version?
2. If they havnt, have they been at least documented somewhere?
|
non_process
|
bad crypto implementation i just saw the comments in this have the bad cryptography designs been addressed in the latest version if they havnt have they been at least documented somewhere
| 0
|
20,521
| 27,180,326,531
|
IssuesEvent
|
2023-02-18 14:48:36
|
OpenDataScotland/the_od_bods
|
https://api.github.com/repos/OpenDataScotland/the_od_bods
|
opened
|
Create alternative .csv output of dataset listing for end users
|
good first issue data processing front end
|
**Is your feature request related to a problem? Please describe.**
Currently, we offer our dataset listing in a .csv and .json download. The .csv was the original output of merge_data.py but is now being replaced by a .json file itself. This means we lose the .csv format for public users.
**Describe the solution you'd like**
Create and a .csv format of the dataset listing that users can download.
**Describe alternatives you've considered**
We've not considered where the .csv should live, so this is up for discussion.
**Additional context**
Original ticket triggering this change is #163
|
1.0
|
Create alternative .csv output of dataset listing for end users - **Is your feature request related to a problem? Please describe.**
Currently, we offer our dataset listing in a .csv and .json download. The .csv was the original output of merge_data.py but is now being replaced by a .json file itself. This means we lose the .csv format for public users.
**Describe the solution you'd like**
Create and a .csv format of the dataset listing that users can download.
**Describe alternatives you've considered**
We've not considered where the .csv should live, so this is up for discussion.
**Additional context**
Original ticket triggering this change is #163
|
process
|
create alternative csv output of dataset listing for end users is your feature request related to a problem please describe currently we offer our dataset listing in a csv and json download the csv was the original output of merge data py but is now being replaced by a json file itself this means we lose the csv format for public users describe the solution you d like create and a csv format of the dataset listing that users can download describe alternatives you ve considered we ve not considered where the csv should live so this is up for discussion additional context original ticket triggering this change is
| 1
|
231,045
| 17,661,016,988
|
IssuesEvent
|
2021-08-21 14:02:12
|
borgbackup/borg
|
https://api.github.com/repos/borgbackup/borg
|
closed
|
borg list patterns doc / example
|
documentation
|
I was trying to use `borg list` with path patterns but the documentation was lacking. I didn't see how to do it until I found a email thread with an example. The docs only say:
> | PATH | paths to list; patterns are supported
My understanding is that unlike with `--exclude`, patterns are not recognized by default, but require a prefix like `re:`. It would be nice to have at least one example and maybe a line that says something like:
> if `path` starts with a pattern prefix, it will be treated like a pattern. Otherwise it must be an exact match."
|
1.0
|
borg list patterns doc / example - I was trying to use `borg list` with path patterns but the documentation was lacking. I didn't see how to do it until I found a email thread with an example. The docs only say:
> | PATH | paths to list; patterns are supported
My understanding is that unlike with `--exclude`, patterns are not recognized by default, but require a prefix like `re:`. It would be nice to have at least one example and maybe a line that says something like:
> if `path` starts with a pattern prefix, it will be treated like a pattern. Otherwise it must be an exact match."
|
non_process
|
borg list patterns doc example i was trying to use borg list with path patterns but the documentation was lacking i didn t see how to do it until i found a email thread with an example the docs only say path paths to list patterns are supported my understanding is that unlike with exclude patterns are not recognized by default but require a prefix like re it would be nice to have at least one example and maybe a line that says something like if path starts with a pattern prefix it will be treated like a pattern otherwise it must be an exact match
| 0
|
170,189
| 13,177,646,772
|
IssuesEvent
|
2020-08-12 07:46:00
|
microsoft/AzureStorageExplorer
|
https://api.github.com/repos/microsoft/AzureStorageExplorer
|
closed
|
No error information displays when typing other service's URL in 'Connect to public blob container' dialog
|
:gear: blobs 🧪 testing
|
**Storage Explorer Version:** 1.15.0-dev
**Build**: 20200717.1
**Branch**: master
**Platform/OS:** Windows 10/ Linux Ubuntu 18.04/ macOS Catalina
**Architecture**: ia32/x64
**Regression From:** Not a regression
**Steps to reproduce:**
1. Expand one storage account -> Copy the URL of one queue.
2. Open connect dialog -> Select 'Connect to public blob container' -> Paste the URL to 'Container URL' box in the dialog.
3. Check the result.
**Expect Experience:**
An error information displays.

**Actual Experience:**
No error information displays.

**More Info:**
This issue also reproduces for file shares/tables.
|
1.0
|
No error information displays when typing other service's URL in 'Connect to public blob container' dialog - **Storage Explorer Version:** 1.15.0-dev
**Build**: 20200717.1
**Branch**: master
**Platform/OS:** Windows 10/ Linux Ubuntu 18.04/ macOS Catalina
**Architecture**: ia32/x64
**Regression From:** Not a regression
**Steps to reproduce:**
1. Expand one storage account -> Copy the URL of one queue.
2. Open connect dialog -> Select 'Connect to public blob container' -> Paste the URL to 'Container URL' box in the dialog.
3. Check the result.
**Expect Experience:**
An error information displays.

**Actual Experience:**
No error information displays.

**More Info:**
This issue also reproduces for file shares/tables.
|
non_process
|
no error information displays when typing other service s url in connect to public blob container dialog storage explorer version dev build branch master platform os windows linux ubuntu macos catalina architecture regression from not a regression steps to reproduce expand one storage account copy the url of one queue open connect dialog select connect to public blob container paste the url to container url box in the dialog check the result expect experience an error information displays actual experience no error information displays more info this issue also reproduces for file shares tables
| 0
|
7,176
| 10,318,672,013
|
IssuesEvent
|
2019-08-30 15:28:54
|
prisma/prisma2
|
https://api.github.com/repos/prisma/prisma2
|
closed
|
Examples and Docs should not use both @id and @unique
|
kind/docs process/candidate
|
In a lot of bug reports i see `@id` and `@unique` being used in conjunction e.g.:
```
model User {
id String @default(cuid()) @id @unique
...
}
```
This is not necessary though. `@id` already implies that something is unique. Hence `@unique` is not needed. I think we should go through our docs and examples and remove it to avoid that this pattern proliferates further.
|
1.0
|
Examples and Docs should not use both @id and @unique - In a lot of bug reports i see `@id` and `@unique` being used in conjunction e.g.:
```
model User {
id String @default(cuid()) @id @unique
...
}
```
This is not necessary though. `@id` already implies that something is unique. Hence `@unique` is not needed. I think we should go through our docs and examples and remove it to avoid that this pattern proliferates further.
|
process
|
examples and docs should not use both id and unique in a lot of bug reports i see id and unique being used in conjunction e g model user id string default cuid id unique this is not necessary though id already implies that something is unique hence unique is not needed i think we should go through our docs and examples and remove it to avoid that this pattern proliferates further
| 1
|
398,193
| 11,739,063,328
|
IssuesEvent
|
2020-03-11 17:04:18
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
closed
|
Cannot use mixed-type question filters on dashboard
|
Priority:P2 Querying/Parameters & Variables Reporting/Dashboards Type:Bug
|
**Describe the bug**
When mixing and matching questions with different filter types (e.g. Field Filter vs. Number/Text Filter), I occasionally receive errors depending on how I load the dashboard.
**Logs**
```
02-10 18:27:49 WARN middleware.process-userland-query :: Query failure {:status :failed,
:class clojure.lang.ExceptionInfo,
:error "Output of value->number does not match schema: \n\n\t (not (matches-some-precondition? nil)) \n\n",
:stacktrace
("--> driver.common.parameters.values$fn__72804$query__GT_params_map__72809$fn__72813.invoke(values.clj:235)"
"driver.common.parameters.values$fn__72804$query__GT_params_map__72809.invoke(values.clj:220)"
"driver.sql$fn__74176.invokeStatic(sql.clj:42)"
"driver.sql$fn__74176.invoke(sql.clj:38)"
"query_processor.middleware.parameters.native$expand_inner.invokeStatic(native.clj:39)"
"query_processor.middleware.parameters.native$expand_inner.invoke(native.clj:30)"
"query_processor.middleware.parameters$expand_one.invokeStatic(parameters.clj:50)"
"query_processor.middleware.parameters$expand_one.invoke(parameters.clj:41)"
"query_processor.middleware.parameters$expand_all$replace_44425__44426.invoke(parameters.clj:59)"
"mbql.util.match$replace_in_collection$iter__26259__26263$fn__26264.invoke(match.clj:132)"
"mbql.util.match$replace_in_collection.invokeStatic(match.clj:131)"
"mbql.util.match$replace_in_collection.invoke(match.clj:126)"
"query_processor.middleware.parameters$expand_all$replace_44425__44426.invoke(parameters.clj:59)"
"query_processor.middleware.parameters$expand_all.invokeStatic(parameters.clj:59)"
"query_processor.middleware.parameters$expand_all.invoke(parameters.clj:53)"
"query_processor.middleware.parameters$expand_all.invokeStatic(parameters.clj:56)"
"query_processor.middleware.parameters$expand_all.invoke(parameters.clj:53)"
"query_processor.middleware.parameters$expand_parameters.invokeStatic(parameters.clj:77)"
"query_processor.middleware.parameters$expand_parameters.invoke(parameters.clj:73)"
"query_processor.middleware.parameters$fn__44441$substitute_parameters_STAR___44446$fn__44447.invoke(parameters.clj:82)"
"query_processor.middleware.parameters$fn__44441$substitute_parameters_STAR___44446.invoke(parameters.clj:79)"
"query_processor.middleware.driver_specific$process_query_in_context$fn__43375.invoke(driver_specific.clj:12)"
"query_processor.middleware.resolve_driver$resolve_driver$fn__44774.invoke(resolve_driver.clj:22)"
"query_processor.middleware.store$initialize_store$fn__47900$fn__47901.invoke(store.clj:11)"
"query_processor.store$do_with_store.invokeStatic(store.clj:46)"
"query_processor.store$do_with_store.invoke(store.clj:40)"
"query_processor.middleware.store$initialize_store$fn__47900.invoke(store.clj:10)"
"query_processor.middleware.async$async__GT_sync$fn__40271.invoke(async.clj:23)"
"query_processor.middleware.async_wait$runnable$fn__42416.invoke(async_wait.clj:89)"),
:query
{:constraints {:max-results 10000, :max-results-bare-rows 2000},
:type :native,
:middleware {:userland-query? true},
:native
{:query "SELECT * FROM ORDERS WHERE ID = {{seller_id}}",
:template-tags {"seller_id" {:id "c26ea0ea-4972-fabc-859b-919b2120926a", :name "seller_id", :display-name "Seller ID", :type :number, :default "2", :required true}}},
:info
{:executed-by 1,
:context :question,
:card-id 28,
:dashboard-id nil,
:query-hash [-89, 90, -6, 117, -80, 81, -118, 122, 11, -32, 121, -50, -122, 70, 40, 79, 70, 9, -5, 69, 40, 61, 122, 33, 10, 113, -83, -112, -66, -41, -70, 28]},
:parameters [{:type "category", :target ["variable" ["template-tag" "seller_id"]], :value ["2" "3"]}],
:async? true,
:cache-ttl nil},
:cause {:class clojure.lang.ExceptionInfo, :error (not (matches-some-precondition? nil)), :ex-data {:type :schema.core/error, :value nil, :error (not (matches-some-precondition? nil))}},
:ex-data
{:type :schema.core/error,
:tags {"seller_id" {:id "c26ea0ea-4972-fabc-859b-919b2120926a", :name "seller_id", :display-name "Seller ID", :type :number, :default "2", :required true}},
:params [{:type :category, :target [:variable [:template-tag "seller_id"]], :value ["2" "3"]}]}}
```
**To Reproduce**
There are two different scenarios where I've seen this happen.
1. Mixed simple/native filter with text/number types
2. Mixed simple/native filter both with field filter types (Postgres)
This issue is for the first, I'll file a second issue for the other.
Steps to reproduce the behavior:
1. Create 3 new questions with sample dataset
2. 1: Simple question, summarize Orders
3. 2: Native question (`SELECT * FROM ORDERS WHERE {{seller_id}}`) with Field filter on Orders-->ID
4. 3. Native question (`SELECT * FROM ORDERS WHERE ID = {{seller_id}}`) with Number filter on Orders-->ID
5. Create a dashboard with those 3 three questions and a filter with a default value
6. Load the dashboard and notice that the native number question doesn't load
re: step 6, sometimes depending on how you load the dashboard, it will work. If I browse to the dashboard in the app, it fails. But if I then reload that page, it loads successfully. If I add a second filter parameter (e.g. http://localhost:3000/dashboard/1?id=2&id=3 ), then it usually always fails.
**Expected behavior**
Questions on the dashboard load regardless of what type of filter I'm using.
**Screenshots**
n/a
**Information about your Metabase Installation:**
```json
{
"browser-info": {
"language": "en-US",
"platform": "MacIntel",
"userAgent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_14_6) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/80.0.3987.87 Safari/537.36",
"vendor": "Google Inc."
},
"system-info": {
"java.runtime.name": "OpenJDK Runtime Environment",
"java.runtime.version": "11.0.4+11",
"java.vendor": "AdoptOpenJDK",
"java.vendor.url": "https://adoptopenjdk.net/",
"java.version": "11.0.4",
"java.vm.name": "OpenJDK 64-Bit Server VM",
"java.vm.version": "11.0.4+11",
"os.name": "Linux",
"os.version": "4.19.76-linuxkit",
"user.language": "en",
"user.timezone": "Etc/UTC"
},
"metabase-info": {
"databases": [
"postgres",
"h2"
],
"hosting-env": "unknown",
"application-database": "postgres",
"run-mode": "prod",
"version": {
"date": "2020-01-16",
"tag": "v1.34.1",
"branch": "enterprise-release-1.34.x",
"hash": "a6df4f6"
},
"settings": {
"report-timezone": null
}
}
}
```
**Severity**
Annoying - I can fix it with a page reload in some cases.
**Additional context**
n/a
|
1.0
|
Cannot use mixed-type question filters on dashboard - **Describe the bug**
When mixing and matching questions with different filter types (e.g. Field Filter vs. Number/Text Filter), I occasionally receive errors depending on how I load the dashboard.
**Logs**
```
02-10 18:27:49 WARN middleware.process-userland-query :: Query failure {:status :failed,
:class clojure.lang.ExceptionInfo,
:error "Output of value->number does not match schema: \n\n\t (not (matches-some-precondition? nil)) \n\n",
:stacktrace
("--> driver.common.parameters.values$fn__72804$query__GT_params_map__72809$fn__72813.invoke(values.clj:235)"
"driver.common.parameters.values$fn__72804$query__GT_params_map__72809.invoke(values.clj:220)"
"driver.sql$fn__74176.invokeStatic(sql.clj:42)"
"driver.sql$fn__74176.invoke(sql.clj:38)"
"query_processor.middleware.parameters.native$expand_inner.invokeStatic(native.clj:39)"
"query_processor.middleware.parameters.native$expand_inner.invoke(native.clj:30)"
"query_processor.middleware.parameters$expand_one.invokeStatic(parameters.clj:50)"
"query_processor.middleware.parameters$expand_one.invoke(parameters.clj:41)"
"query_processor.middleware.parameters$expand_all$replace_44425__44426.invoke(parameters.clj:59)"
"mbql.util.match$replace_in_collection$iter__26259__26263$fn__26264.invoke(match.clj:132)"
"mbql.util.match$replace_in_collection.invokeStatic(match.clj:131)"
"mbql.util.match$replace_in_collection.invoke(match.clj:126)"
"query_processor.middleware.parameters$expand_all$replace_44425__44426.invoke(parameters.clj:59)"
"query_processor.middleware.parameters$expand_all.invokeStatic(parameters.clj:59)"
"query_processor.middleware.parameters$expand_all.invoke(parameters.clj:53)"
"query_processor.middleware.parameters$expand_all.invokeStatic(parameters.clj:56)"
"query_processor.middleware.parameters$expand_all.invoke(parameters.clj:53)"
"query_processor.middleware.parameters$expand_parameters.invokeStatic(parameters.clj:77)"
"query_processor.middleware.parameters$expand_parameters.invoke(parameters.clj:73)"
"query_processor.middleware.parameters$fn__44441$substitute_parameters_STAR___44446$fn__44447.invoke(parameters.clj:82)"
"query_processor.middleware.parameters$fn__44441$substitute_parameters_STAR___44446.invoke(parameters.clj:79)"
"query_processor.middleware.driver_specific$process_query_in_context$fn__43375.invoke(driver_specific.clj:12)"
"query_processor.middleware.resolve_driver$resolve_driver$fn__44774.invoke(resolve_driver.clj:22)"
"query_processor.middleware.store$initialize_store$fn__47900$fn__47901.invoke(store.clj:11)"
"query_processor.store$do_with_store.invokeStatic(store.clj:46)"
"query_processor.store$do_with_store.invoke(store.clj:40)"
"query_processor.middleware.store$initialize_store$fn__47900.invoke(store.clj:10)"
"query_processor.middleware.async$async__GT_sync$fn__40271.invoke(async.clj:23)"
"query_processor.middleware.async_wait$runnable$fn__42416.invoke(async_wait.clj:89)"),
:query
{:constraints {:max-results 10000, :max-results-bare-rows 2000},
:type :native,
:middleware {:userland-query? true},
:native
{:query "SELECT * FROM ORDERS WHERE ID = {{seller_id}}",
:template-tags {"seller_id" {:id "c26ea0ea-4972-fabc-859b-919b2120926a", :name "seller_id", :display-name "Seller ID", :type :number, :default "2", :required true}}},
:info
{:executed-by 1,
:context :question,
:card-id 28,
:dashboard-id nil,
:query-hash [-89, 90, -6, 117, -80, 81, -118, 122, 11, -32, 121, -50, -122, 70, 40, 79, 70, 9, -5, 69, 40, 61, 122, 33, 10, 113, -83, -112, -66, -41, -70, 28]},
:parameters [{:type "category", :target ["variable" ["template-tag" "seller_id"]], :value ["2" "3"]}],
:async? true,
:cache-ttl nil},
:cause {:class clojure.lang.ExceptionInfo, :error (not (matches-some-precondition? nil)), :ex-data {:type :schema.core/error, :value nil, :error (not (matches-some-precondition? nil))}},
:ex-data
{:type :schema.core/error,
:tags {"seller_id" {:id "c26ea0ea-4972-fabc-859b-919b2120926a", :name "seller_id", :display-name "Seller ID", :type :number, :default "2", :required true}},
:params [{:type :category, :target [:variable [:template-tag "seller_id"]], :value ["2" "3"]}]}}
```
**To Reproduce**
There are two different scenarios where I've seen this happen.
1. Mixed simple/native filter with text/number types
2. Mixed simple/native filter both with field filter types (Postgres)
This issue is for the first, I'll file a second issue for the other.
Steps to reproduce the behavior:
1. Create 3 new questions with sample dataset
2. 1: Simple question, summarize Orders
3. 2: Native question (`SELECT * FROM ORDERS WHERE {{seller_id}}`) with Field filter on Orders-->ID
4. 3. Native question (`SELECT * FROM ORDERS WHERE ID = {{seller_id}}`) with Number filter on Orders-->ID
5. Create a dashboard with those 3 three questions and a filter with a default value
6. Load the dashboard and notice that the native number question doesn't load
re: step 6, sometimes depending on how you load the dashboard, it will work. If I browse to the dashboard in the app, it fails. But if I then reload that page, it loads successfully. If I add a second filter parameter (e.g. http://localhost:3000/dashboard/1?id=2&id=3 ), then it usually always fails.
**Expected behavior**
Questions on the dashboard load regardless of what type of filter I'm using.
**Screenshots**
n/a
**Information about your Metabase Installation:**
```json
{
"browser-info": {
"language": "en-US",
"platform": "MacIntel",
"userAgent": "Mozilla/5.0 (Macintosh; Intel Mac OS X 10_14_6) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/80.0.3987.87 Safari/537.36",
"vendor": "Google Inc."
},
"system-info": {
"java.runtime.name": "OpenJDK Runtime Environment",
"java.runtime.version": "11.0.4+11",
"java.vendor": "AdoptOpenJDK",
"java.vendor.url": "https://adoptopenjdk.net/",
"java.version": "11.0.4",
"java.vm.name": "OpenJDK 64-Bit Server VM",
"java.vm.version": "11.0.4+11",
"os.name": "Linux",
"os.version": "4.19.76-linuxkit",
"user.language": "en",
"user.timezone": "Etc/UTC"
},
"metabase-info": {
"databases": [
"postgres",
"h2"
],
"hosting-env": "unknown",
"application-database": "postgres",
"run-mode": "prod",
"version": {
"date": "2020-01-16",
"tag": "v1.34.1",
"branch": "enterprise-release-1.34.x",
"hash": "a6df4f6"
},
"settings": {
"report-timezone": null
}
}
}
```
**Severity**
Annoying - I can fix it with a page reload in some cases.
**Additional context**
n/a
|
non_process
|
cannot use mixed type question filters on dashboard describe the bug when mixing and matching questions with different filter types e g field filter vs number text filter i occasionally receive errors depending on how i load the dashboard logs warn middleware process userland query query failure status failed class clojure lang exceptioninfo error output of value number does not match schema n n t not matches some precondition nil n n stacktrace driver common parameters values fn query gt params map fn invoke values clj driver common parameters values fn query gt params map invoke values clj driver sql fn invokestatic sql clj driver sql fn invoke sql clj query processor middleware parameters native expand inner invokestatic native clj query processor middleware parameters native expand inner invoke native clj query processor middleware parameters expand one invokestatic parameters clj query processor middleware parameters expand one invoke parameters clj query processor middleware parameters expand all replace invoke parameters clj mbql util match replace in collection iter fn invoke match clj mbql util match replace in collection invokestatic match clj mbql util match replace in collection invoke match clj query processor middleware parameters expand all replace invoke parameters clj query processor middleware parameters expand all invokestatic parameters clj query processor middleware parameters expand all invoke parameters clj query processor middleware parameters expand all invokestatic parameters clj query processor middleware parameters expand all invoke parameters clj query processor middleware parameters expand parameters invokestatic parameters clj query processor middleware parameters expand parameters invoke parameters clj query processor middleware parameters fn substitute parameters star fn invoke parameters clj query processor middleware parameters fn substitute parameters star invoke parameters clj query processor middleware driver specific process query in context fn invoke driver specific clj query processor middleware resolve driver resolve driver fn invoke resolve driver clj query processor middleware store initialize store fn fn invoke store clj query processor store do with store invokestatic store clj query processor store do with store invoke store clj query processor middleware store initialize store fn invoke store clj query processor middleware async async gt sync fn invoke async clj query processor middleware async wait runnable fn invoke async wait clj query constraints max results max results bare rows type native middleware userland query true native query select from orders where id seller id template tags seller id id fabc name seller id display name seller id type number default required true info executed by context question card id dashboard id nil query hash parameters value async true cache ttl nil cause class clojure lang exceptioninfo error not matches some precondition nil ex data type schema core error value nil error not matches some precondition nil ex data type schema core error tags seller id id fabc name seller id display name seller id type number default required true params value to reproduce there are two different scenarios where i ve seen this happen mixed simple native filter with text number types mixed simple native filter both with field filter types postgres this issue is for the first i ll file a second issue for the other steps to reproduce the behavior create new questions with sample dataset simple question summarize orders native question select from orders where seller id with field filter on orders id native question select from orders where id seller id with number filter on orders id create a dashboard with those three questions and a filter with a default value load the dashboard and notice that the native number question doesn t load re step sometimes depending on how you load the dashboard it will work if i browse to the dashboard in the app it fails but if i then reload that page it loads successfully if i add a second filter parameter e g then it usually always fails expected behavior questions on the dashboard load regardless of what type of filter i m using screenshots n a information about your metabase installation json browser info language en us platform macintel useragent mozilla macintosh intel mac os x applewebkit khtml like gecko chrome safari vendor google inc system info java runtime name openjdk runtime environment java runtime version java vendor adoptopenjdk java vendor url java version java vm name openjdk bit server vm java vm version os name linux os version linuxkit user language en user timezone etc utc metabase info databases postgres hosting env unknown application database postgres run mode prod version date tag branch enterprise release x hash settings report timezone null severity annoying i can fix it with a page reload in some cases additional context n a
| 0
|
280,339
| 24,296,338,240
|
IssuesEvent
|
2022-09-29 10:21:31
|
wpfoodmanager/wp-food-manager
|
https://api.github.com/repos/wpfoodmanager/wp-food-manager
|
closed
|
Backend - Able to change food postion
|
In Testing
|
Able to change food postion . This should not be change.
https://user-images.githubusercontent.com/75515088/192772946-8ff3c4b0-f19d-4f38-83f1-d7b8c5e0c6f5.mp4
|
1.0
|
Backend - Able to change food postion - Able to change food postion . This should not be change.
https://user-images.githubusercontent.com/75515088/192772946-8ff3c4b0-f19d-4f38-83f1-d7b8c5e0c6f5.mp4
|
non_process
|
backend able to change food postion able to change food postion this should not be change
| 0
|
23,740
| 16,550,549,680
|
IssuesEvent
|
2021-05-28 08:05:14
|
google/site-kit-wp
|
https://api.github.com/repos/google/site-kit-wp
|
opened
|
Add a security policy
|
P1 Type: Infrastructure
|
## Feature Description
https://docs.github.com/en/code-security/getting-started/adding-a-security-policy-to-your-repository
Eg. https://github.com/ampproject/amp-wp/blob/3bb15f3d660c4401b958147430664d07eb1e03e2/SECURITY.md
---------------
_Do not alter or remove anything below. The following sections will be managed by moderators only._
## Acceptance criteria
* <!-- One or more bullet points for acceptance criteria. -->
## Implementation Brief
* <!-- One or more bullet points for how to technically implement the feature. -->
### Test Coverage
* <!-- One or more bullet points for how to implement automated tests to verify the feature works. -->
### Visual Regression Changes
* <!-- One or more bullet points describing how the feature will affect visual regression tests, if applicable. -->
## QA Brief
* <!-- One or more bullet points for how to test that the feature works as expected. -->
## Changelog entry
* <!-- One sentence summarizing the PR, to be used in the changelog. -->
|
1.0
|
Add a security policy - ## Feature Description
https://docs.github.com/en/code-security/getting-started/adding-a-security-policy-to-your-repository
Eg. https://github.com/ampproject/amp-wp/blob/3bb15f3d660c4401b958147430664d07eb1e03e2/SECURITY.md
---------------
_Do not alter or remove anything below. The following sections will be managed by moderators only._
## Acceptance criteria
* <!-- One or more bullet points for acceptance criteria. -->
## Implementation Brief
* <!-- One or more bullet points for how to technically implement the feature. -->
### Test Coverage
* <!-- One or more bullet points for how to implement automated tests to verify the feature works. -->
### Visual Regression Changes
* <!-- One or more bullet points describing how the feature will affect visual regression tests, if applicable. -->
## QA Brief
* <!-- One or more bullet points for how to test that the feature works as expected. -->
## Changelog entry
* <!-- One sentence summarizing the PR, to be used in the changelog. -->
|
non_process
|
add a security policy feature description eg do not alter or remove anything below the following sections will be managed by moderators only acceptance criteria implementation brief test coverage visual regression changes qa brief changelog entry
| 0
|
40,949
| 10,238,386,188
|
IssuesEvent
|
2019-08-19 15:46:16
|
PowerDNS/pdns
|
https://api.github.com/repos/PowerDNS/pdns
|
opened
|
auth nsupdate: duplicate entries when mixing case
|
auth defect
|
- Program: Authoritative
- Issue type: Bug report/Feature request
### Short description
Our 2136-code prevents adding duplicate entries, but this fails in the face of mixed case entries.
Patching one of the existing tests like this reveals the issue:
```diff
diff --git a/regression-tests/tests/1dyndns-update-replace-cname/command b/regression-tests/tests/1dyndns-update-replace-cname/command
index 17bd7c818..3eef41f0e 100755
--- a/regression-tests/tests/1dyndns-update-replace-cname/command
+++ b/regression-tests/tests/1dyndns-update-replace-cname/command
@@ -1,22 +1,22 @@
#!/bin/sh
-cleandig cname1.test.dyndns CNAME
+cleandig ptr1.test.dyndns PTR
cleannsupdate <<!
server $nameserver $port
zone test.dyndns
-update add cname1.test.dyndns. 3600 CNAME host-2.test.dyndns.
+update add ptr1.test.dyndns. 3600 PTR host-2.test.dyndns.
send
answer
!
-cleandig cname1.test.dyndns CNAME
+cleandig ptr1.test.dyndns PTR
cleannsupdate <<!
server $nameserver $port
zone test.dyndns
-update add cname1.test.dyndns. 3600 CNAME host-1.test.dyndns.
+update add ptr1.test.dyndns. 3600 PTR HOST-2.test.dyndns.
send
answer
!
-cleandig cname1.test.dyndns CNAME
+cleandig ptr1.test.dyndns PTR
```
|
1.0
|
auth nsupdate: duplicate entries when mixing case - - Program: Authoritative
- Issue type: Bug report/Feature request
### Short description
Our 2136-code prevents adding duplicate entries, but this fails in the face of mixed case entries.
Patching one of the existing tests like this reveals the issue:
```diff
diff --git a/regression-tests/tests/1dyndns-update-replace-cname/command b/regression-tests/tests/1dyndns-update-replace-cname/command
index 17bd7c818..3eef41f0e 100755
--- a/regression-tests/tests/1dyndns-update-replace-cname/command
+++ b/regression-tests/tests/1dyndns-update-replace-cname/command
@@ -1,22 +1,22 @@
#!/bin/sh
-cleandig cname1.test.dyndns CNAME
+cleandig ptr1.test.dyndns PTR
cleannsupdate <<!
server $nameserver $port
zone test.dyndns
-update add cname1.test.dyndns. 3600 CNAME host-2.test.dyndns.
+update add ptr1.test.dyndns. 3600 PTR host-2.test.dyndns.
send
answer
!
-cleandig cname1.test.dyndns CNAME
+cleandig ptr1.test.dyndns PTR
cleannsupdate <<!
server $nameserver $port
zone test.dyndns
-update add cname1.test.dyndns. 3600 CNAME host-1.test.dyndns.
+update add ptr1.test.dyndns. 3600 PTR HOST-2.test.dyndns.
send
answer
!
-cleandig cname1.test.dyndns CNAME
+cleandig ptr1.test.dyndns PTR
```
|
non_process
|
auth nsupdate duplicate entries when mixing case program authoritative issue type bug report feature request short description our code prevents adding duplicate entries but this fails in the face of mixed case entries patching one of the existing tests like this reveals the issue diff diff git a regression tests tests update replace cname command b regression tests tests update replace cname command index a regression tests tests update replace cname command b regression tests tests update replace cname command bin sh cleandig test dyndns cname cleandig test dyndns ptr cleannsupdate server nameserver port zone test dyndns update add test dyndns cname host test dyndns update add test dyndns ptr host test dyndns send answer cleandig test dyndns cname cleandig test dyndns ptr cleannsupdate server nameserver port zone test dyndns update add test dyndns cname host test dyndns update add test dyndns ptr host test dyndns send answer cleandig test dyndns cname cleandig test dyndns ptr
| 0
|
6,182
| 9,100,315,465
|
IssuesEvent
|
2019-02-20 08:11:00
|
comic/grand-challenge.org
|
https://api.github.com/repos/comic/grand-challenge.org
|
opened
|
Replace mhd/zraw with mha as internal image representation?
|
area/ophthalmology-workstation area/processors
|
I think that we should replace mhd/zraw with compressed mha as our internal representation. I think that this would have the following advantages:
- At the moment, we re-write all meta image files to be named out.mhd and out.zraw, which are located at `images/<image.pk>/out.<type>`. This is awkward as everywhere we need to handle two files, sort out which one is the header, and already in the codebase there are a lot of select by file extension statements.
- If we want to share a bunch of files, say, as an ImageSet for training, all of the files will have the same name, and will have to be re-written (including mhd modification) before they're given to the user.
I would prefer that we use mha files. Then, we only have to deal with 1 file, we don't need to sort it everywhere, and we can rename it on the fly. This would then be consistent with the tiff representation. We could then serve images from `images/<image.pk>/<imagefile.pk>/`.
I'm fairly sure that this would still be compatible with the Python (SimpleITK), Mevislab and Javascript (ITK.js) loading libraries.
What could break:
- Everywhere that we've made an assumption about getting 2 images:
- CIRRUS Web
- The `get_sitk_image` function (mhd is not used directly by the ophthalmology workstation).
- Processors: how do the existing algorithms handle mhd?
Any thoughts? cc. @pkcakeout @HarmvZ
|
1.0
|
Replace mhd/zraw with mha as internal image representation? - I think that we should replace mhd/zraw with compressed mha as our internal representation. I think that this would have the following advantages:
- At the moment, we re-write all meta image files to be named out.mhd and out.zraw, which are located at `images/<image.pk>/out.<type>`. This is awkward as everywhere we need to handle two files, sort out which one is the header, and already in the codebase there are a lot of select by file extension statements.
- If we want to share a bunch of files, say, as an ImageSet for training, all of the files will have the same name, and will have to be re-written (including mhd modification) before they're given to the user.
I would prefer that we use mha files. Then, we only have to deal with 1 file, we don't need to sort it everywhere, and we can rename it on the fly. This would then be consistent with the tiff representation. We could then serve images from `images/<image.pk>/<imagefile.pk>/`.
I'm fairly sure that this would still be compatible with the Python (SimpleITK), Mevislab and Javascript (ITK.js) loading libraries.
What could break:
- Everywhere that we've made an assumption about getting 2 images:
- CIRRUS Web
- The `get_sitk_image` function (mhd is not used directly by the ophthalmology workstation).
- Processors: how do the existing algorithms handle mhd?
Any thoughts? cc. @pkcakeout @HarmvZ
|
process
|
replace mhd zraw with mha as internal image representation i think that we should replace mhd zraw with compressed mha as our internal representation i think that this would have the following advantages at the moment we re write all meta image files to be named out mhd and out zraw which are located at images out this is awkward as everywhere we need to handle two files sort out which one is the header and already in the codebase there are a lot of select by file extension statements if we want to share a bunch of files say as an imageset for training all of the files will have the same name and will have to be re written including mhd modification before they re given to the user i would prefer that we use mha files then we only have to deal with file we don t need to sort it everywhere and we can rename it on the fly this would then be consistent with the tiff representation we could then serve images from images i m fairly sure that this would still be compatible with the python simpleitk mevislab and javascript itk js loading libraries what could break everywhere that we ve made an assumption about getting images cirrus web the get sitk image function mhd is not used directly by the ophthalmology workstation processors how do the existing algorithms handle mhd any thoughts cc pkcakeout harmvz
| 1
|
14,546
| 17,662,984,241
|
IssuesEvent
|
2021-08-21 22:32:52
|
GSG-FC03/adnan-Tic-Tac-Toe
|
https://api.github.com/repos/GSG-FC03/adnan-Tic-Tac-Toe
|
opened
|
functionality
|
in-process T 5hr
|
Creating functions that make the Tic Tac Toe works
- [ ] Place the mark
- [ ] check for winner
- [ ] check for draw
- [ ] switch turns
|
1.0
|
functionality - Creating functions that make the Tic Tac Toe works
- [ ] Place the mark
- [ ] check for winner
- [ ] check for draw
- [ ] switch turns
|
process
|
functionality creating functions that make the tic tac toe works place the mark check for winner check for draw switch turns
| 1
|
1,151
| 3,066,874,863
|
IssuesEvent
|
2015-08-18 06:32:23
|
TeamMentor/TM_4_0_Design
|
https://api.github.com/repos/TeamMentor/TM_4_0_Design
|
closed
|
Map out current security practices with 'Simplified Implementation of the Microsoft SDL'
|
Area: Security P3 Type: Task
|
* http://www.microsoft.com/security/sdl/default.aspx
* docs can be downloaded from http://www.microsoft.com/en-us/download/details.aspx?id=12379

We are already doing a lot of these, but it will be good to formalize them (for example on the one SDL steps that is missing is [Create a Threat Model for TM 4.0 (Jade, Flare and GraphDB)](https://github.com/TeamMentor/TM_4_0_Design/issues/269) )
----
[TM-4.0-Security](https://github.com/TeamMentor/TM_4_0_Design/wiki/TM-4.0-Security)
|
True
|
Map out current security practices with 'Simplified Implementation of the Microsoft SDL' - * http://www.microsoft.com/security/sdl/default.aspx
* docs can be downloaded from http://www.microsoft.com/en-us/download/details.aspx?id=12379

We are already doing a lot of these, but it will be good to formalize them (for example on the one SDL steps that is missing is [Create a Threat Model for TM 4.0 (Jade, Flare and GraphDB)](https://github.com/TeamMentor/TM_4_0_Design/issues/269) )
----
[TM-4.0-Security](https://github.com/TeamMentor/TM_4_0_Design/wiki/TM-4.0-Security)
|
non_process
|
map out current security practices with simplified implementation of the microsoft sdl docs can be downloaded from we are already doing a lot of these but it will be good to formalize them for example on the one sdl steps that is missing is
| 0
|
37,475
| 8,301,518,280
|
IssuesEvent
|
2018-09-21 11:44:44
|
Altinn/altinn-studio
|
https://api.github.com/repos/Altinn/altinn-studio
|
opened
|
As a service developer I should be able to add translations for codelist values
|
codelist
|
Kunne legge inn tekstnøkler
Apiet skal bytte ut tekstnøkler med tekst
|
1.0
|
As a service developer I should be able to add translations for codelist values - Kunne legge inn tekstnøkler
Apiet skal bytte ut tekstnøkler med tekst
|
non_process
|
as a service developer i should be able to add translations for codelist values kunne legge inn tekstnøkler apiet skal bytte ut tekstnøkler med tekst
| 0
|
55,441
| 30,753,416,195
|
IssuesEvent
|
2023-07-28 21:53:47
|
reportportal/reportportal
|
https://api.github.com/repos/reportportal/reportportal
|
closed
|
Unknown 'Auto-analysis' end status.
|
bug Check: Performance
|
**Describe the bug**
1. Open Launches page
2. Open 'More' menu for particular launch
3. Choose 'Analysis'
4. Auto-analysis started
**Actual behavior**
'Auto-analysis' icon appears and constantly flashing all time.
There is no any provided status of analysis. How should I recognise when it finishes?
**Expected behavior**
The end status of auto-analysis is displayed.
**Screenshots**

**Versions:**
- macOS 13.3.1 (a), Chrome 114.0.5735.133
- Version of RP 23.1
|
True
|
Unknown 'Auto-analysis' end status. - **Describe the bug**
1. Open Launches page
2. Open 'More' menu for particular launch
3. Choose 'Analysis'
4. Auto-analysis started
**Actual behavior**
'Auto-analysis' icon appears and constantly flashing all time.
There is no any provided status of analysis. How should I recognise when it finishes?
**Expected behavior**
The end status of auto-analysis is displayed.
**Screenshots**

**Versions:**
- macOS 13.3.1 (a), Chrome 114.0.5735.133
- Version of RP 23.1
|
non_process
|
unknown auto analysis end status describe the bug open launches page open more menu for particular launch choose analysis auto analysis started actual behavior auto analysis icon appears and constantly flashing all time there is no any provided status of analysis how should i recognise when it finishes expected behavior the end status of auto analysis is displayed screenshots versions macos a chrome version of rp
| 0
|
617
| 3,083,671,224
|
IssuesEvent
|
2015-08-24 10:28:51
|
Wikitalia/edgesense
|
https://api.github.com/repos/Wikitalia/edgesense
|
opened
|
Enable analysis to concentrate on last XX months of community
|
enhancement processing
|
It would be useful to have an analysis and visualization for the last period of time (year / six months), in addition to the all time network.
It gives an idea on the status quo and who is active now. Also it is useful for older communities as InnovatoriPA were there are a lot of old nodes from the beginning which are not active any more and the shape of the network could be somehow different now with the only active ones. Also it is needed for activity management and in showing how each activity changed by time.
|
1.0
|
Enable analysis to concentrate on last XX months of community - It would be useful to have an analysis and visualization for the last period of time (year / six months), in addition to the all time network.
It gives an idea on the status quo and who is active now. Also it is useful for older communities as InnovatoriPA were there are a lot of old nodes from the beginning which are not active any more and the shape of the network could be somehow different now with the only active ones. Also it is needed for activity management and in showing how each activity changed by time.
|
process
|
enable analysis to concentrate on last xx months of community it would be useful to have an analysis and visualization for the last period of time year six months in addition to the all time network it gives an idea on the status quo and who is active now also it is useful for older communities as innovatoripa were there are a lot of old nodes from the beginning which are not active any more and the shape of the network could be somehow different now with the only active ones also it is needed for activity management and in showing how each activity changed by time
| 1
|
36,828
| 8,148,946,130
|
IssuesEvent
|
2018-08-22 08:02:34
|
codl/forget
|
https://api.github.com/repos/codl/forget
|
opened
|
support new twitter archive format
|
defect
|
reported by `rrix@cybre.space` <https://cybre.space/@rrix/100591445673683791>
> hey, it looks like the Twitter archive format changed at some point that makes it not work with Forget. there's no longer a data/js/tweets dir with monthly files, just a big jsonp file (67mib in my case) in the root of the zip
|
1.0
|
support new twitter archive format - reported by `rrix@cybre.space` <https://cybre.space/@rrix/100591445673683791>
> hey, it looks like the Twitter archive format changed at some point that makes it not work with Forget. there's no longer a data/js/tweets dir with monthly files, just a big jsonp file (67mib in my case) in the root of the zip
|
non_process
|
support new twitter archive format reported by rrix cybre space hey it looks like the twitter archive format changed at some point that makes it not work with forget there s no longer a data js tweets dir with monthly files just a big jsonp file in my case in the root of the zip
| 0
|
122,876
| 16,372,799,056
|
IssuesEvent
|
2021-05-15 13:42:29
|
TCastus/mobilite2-front
|
https://api.github.com/repos/TCastus/mobilite2-front
|
closed
|
Design de la page d'accueil
|
design
|
Sur Canva, proposer une interface pour la page d'accueil avant l'implémentation
|
1.0
|
Design de la page d'accueil - Sur Canva, proposer une interface pour la page d'accueil avant l'implémentation
|
non_process
|
design de la page d accueil sur canva proposer une interface pour la page d accueil avant l implémentation
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.