Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
71,904
18,923,046,729
IssuesEvent
2021-11-17 05:43:44
ARM-software/armnn
https://api.github.com/repos/ARM-software/armnn
closed
Compile error for TFlite Delegate
Documentation issue Build issue
aim at: **Build the TfLite Delegate (Stand-Alone)** ref from:https://github.com/ARM-software/armnn/blob/branches%2Farmnn_21_02/delegate/BuildGuideNative.md env: **1) ubuntu18.04(x86) 2) tensorflow-2.3.1 -build success 3) flatbuffers-1.12.0-build arm64 success 4) compute library_21.02 -build success** `scons arch=arm64-v8a neon=1 opencl=1 embed_kernels=1 extra_cxx_flags="-fPIC" benchmark_tests=0 validation_tests=0 -j8 internal_only=0` **5)build armnn 21.02 success** ``` CXX=arrch64-linux-gnu-g++ CC=aarch64-linux-gnu-gcc cmake .. -DARMCOMPUTE_ROOT=$BASEDIR/ComputeLibrary -DARMCOMPUTENEON=1 -DBUILD_UNIT_TESTS=0 make ``` **6) build tflte delegate** ``` CXX=arrch64-linux-gnu-g++ CC=aarch64-linux-gnu-gcc cmake .. -DTENSORFLOW_LIB_DIR=$BASEDIR/tensorflow/bazel-bin -DTENSORFLOW_ROOT=$BASEDIR/tensorflow -DTFLITE_LIB_ROOT=$BASEDIR/tensorflow/bazel-bin -DFLATBUFFERS_ROOT=$BASEDIR/flatbuffers/install -DArmnn_DIR=$BASEDIR/armnn/build -DARMNN_SOURCE_DIR=$BASEDIR/armnn make ``` /usr/lib/gcc-cross/aarch64-linux-gnu/7/../../../aarch64-linux-gnu/bin/ld:skipping incompatible /home/***/delegate/tensorflow/bazel-bin/libtensorflow_lite_all.so when searching for -ltensorflow_lite_all /usr/lib/gcc-cross/aarch64-linux-gnu/7/../../../aarch64-linux-gnu/bin/ld: cannot find -ltensorflow_lite_all look at dirctory,libtensorflow_lite_all.so is exist: ``` ls ../../../tensorflow-bazel-bin libtensorflow_lite_all.so ``` please give me some advice to slove it,thanks。
1.0
Compile error for TFlite Delegate - aim at: **Build the TfLite Delegate (Stand-Alone)** ref from:https://github.com/ARM-software/armnn/blob/branches%2Farmnn_21_02/delegate/BuildGuideNative.md env: **1) ubuntu18.04(x86) 2) tensorflow-2.3.1 -build success 3) flatbuffers-1.12.0-build arm64 success 4) compute library_21.02 -build success** `scons arch=arm64-v8a neon=1 opencl=1 embed_kernels=1 extra_cxx_flags="-fPIC" benchmark_tests=0 validation_tests=0 -j8 internal_only=0` **5)build armnn 21.02 success** ``` CXX=arrch64-linux-gnu-g++ CC=aarch64-linux-gnu-gcc cmake .. -DARMCOMPUTE_ROOT=$BASEDIR/ComputeLibrary -DARMCOMPUTENEON=1 -DBUILD_UNIT_TESTS=0 make ``` **6) build tflte delegate** ``` CXX=arrch64-linux-gnu-g++ CC=aarch64-linux-gnu-gcc cmake .. -DTENSORFLOW_LIB_DIR=$BASEDIR/tensorflow/bazel-bin -DTENSORFLOW_ROOT=$BASEDIR/tensorflow -DTFLITE_LIB_ROOT=$BASEDIR/tensorflow/bazel-bin -DFLATBUFFERS_ROOT=$BASEDIR/flatbuffers/install -DArmnn_DIR=$BASEDIR/armnn/build -DARMNN_SOURCE_DIR=$BASEDIR/armnn make ``` /usr/lib/gcc-cross/aarch64-linux-gnu/7/../../../aarch64-linux-gnu/bin/ld:skipping incompatible /home/***/delegate/tensorflow/bazel-bin/libtensorflow_lite_all.so when searching for -ltensorflow_lite_all /usr/lib/gcc-cross/aarch64-linux-gnu/7/../../../aarch64-linux-gnu/bin/ld: cannot find -ltensorflow_lite_all look at dirctory,libtensorflow_lite_all.so is exist: ``` ls ../../../tensorflow-bazel-bin libtensorflow_lite_all.so ``` please give me some advice to slove it,thanks。
non_process
compile error for tflite delegate aim at build the tflite delegate stand alone ref from env tensorflow build success flatbuffers build success compute library build success scons arch neon opencl embed kernels extra cxx flags fpic benchmark tests validation tests internal only build armnn success cxx linux gnu g cc linux gnu gcc cmake darmcompute root basedir computelibrary darmcomputeneon dbuild unit tests make build tflte delegate cxx linux gnu g cc linux gnu gcc cmake dtensorflow lib dir basedir tensorflow bazel bin dtensorflow root basedir tensorflow dtflite lib root basedir tensorflow bazel bin dflatbuffers root basedir flatbuffers install darmnn dir basedir armnn build darmnn source dir basedir armnn make usr lib gcc cross linux gnu linux gnu bin ld skipping incompatible home delegate tensorflow bazel bin libtensorflow lite all so when searching for ltensorflow lite all usr lib gcc cross linux gnu linux gnu bin ld cannot find ltensorflow lite all look at dirctory,libtensorflow lite all so is exist: ls tensorflow bazel bin libtensorflow lite all so please give me some advice to slove it,thanks。
0
8,270
11,430,788,728
IssuesEvent
2020-02-04 10:46:25
Graylog2/graylog2-server
https://api.github.com/repos/Graylog2/graylog2-server
closed
Grok pattern Extractors crashes and drop logs
bug processing
After I upgraded Graylog to 3.2 I found that one extractor containing a grok pattern was constantly crashing, the logs were then forever lost. ## Expected Behavior When I add an extractor with a grok pattern for syslog and select "Always try to extract", graylog should not drop the logs that failed! ## Current Behavior I added the following extractor : ![Screenshot_2020-02-04 Graylog - New extractor for input Beats(1)](https://user-images.githubusercontent.com/31347719/73732090-d7f3ab80-4739-11ea-9260-c2b275a3bf62.png) And I can see in graylog-server logs : ``` 2020-02-04T10:16:09.713+01:00 WARN [ProcessBufferProcessor] Unable to process message <fb393704-472e-11ea-bb88-4256e21dfc8e>: java.lang.ClassCastException: Cannot cast java.lang.String to org.joda.time.DateTime 2020-02-04T10:16:09.713+01:00 WARN [ProcessBufferProcessor] Unable to process message <fb393705-472e-11ea-bb88-4256e21dfc8e>: java.lang.ClassCastException: Cannot cast java.lang.String to org.joda.time.DateTime 2020-02-04T10:16:09.713+01:00 WARN [ProcessBufferProcessor] Unable to process message <fb393706-472e-11ea-bb88-4256e21dfc8e>: java.lang.ClassCastException: Cannot cast java.lang.String to org.joda.time.DateTime 2020-02-04T10:16:09.958+01:00 WARN [ProcessBufferProcessor] Unable to process message <fb5e7245-472e-11ea-bb88-4256e21dfc8e>: java.lang.ClassCastException: Cannot cast java.lang.String to org.joda.time.DateTime 2020-02-04T10:16:09.958+01:00 WARN [ProcessBufferProcessor] Unable to process message <fb5e7244-472e-11ea-bb88-4256e21dfc8e>: java.lang.ClassCastException: Cannot cast java.lang.String to org.joda.time.DateTime 2020-02-04T10:16:09.958+01:00 WARN [ProcessBufferProcessor] Unable to process message <fb5e7243-472e-11ea-bb88-4256e21dfc8e>: java.lang.ClassCastException: Cannot cast java.lang.String to org.joda.time.DateTime 2020-02-04T10:16:10.076+01:00 WARN [ProcessBufferProcessor] Unable to process message <fb709ab4-472e-11ea-bb88-4256e21dfc8e>: java.lang.ClassCastException: Cannot cast java.lang.String to org.joda.time.DateTime 2020-02-04T10:16:10.077+01:00 WARN [ProcessBufferProcessor] Unable to process message <fb70c1c0-472e-11ea-bb88-4256e21dfc8e>: java.lang.ClassCastException: Cannot cast java.lang.String to org.joda.time.DateTime ``` When this happened, I have lost a lot of logs that were not indexed in graylog. When I delete the extractor everything work smoothly. ## Possible Solution Catch the error so that the logs can be indexed even tho the extractor failed ? ## Steps to Reproduce (for bugs) 1. Graylog in 3.2 2. Add an extractor with the Grok Pattern from above 3. Add logs that do not match the Grok Pattern to Graylog ## Context This bug automatically dropped logs without warning. ## Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> * Graylog Version: 3.2 * Elasticsearch Version: * MongoDB Version: * Operating System: Linux/Debian * Browser version: Firefox
1.0
Grok pattern Extractors crashes and drop logs - After I upgraded Graylog to 3.2 I found that one extractor containing a grok pattern was constantly crashing, the logs were then forever lost. ## Expected Behavior When I add an extractor with a grok pattern for syslog and select "Always try to extract", graylog should not drop the logs that failed! ## Current Behavior I added the following extractor : ![Screenshot_2020-02-04 Graylog - New extractor for input Beats(1)](https://user-images.githubusercontent.com/31347719/73732090-d7f3ab80-4739-11ea-9260-c2b275a3bf62.png) And I can see in graylog-server logs : ``` 2020-02-04T10:16:09.713+01:00 WARN [ProcessBufferProcessor] Unable to process message <fb393704-472e-11ea-bb88-4256e21dfc8e>: java.lang.ClassCastException: Cannot cast java.lang.String to org.joda.time.DateTime 2020-02-04T10:16:09.713+01:00 WARN [ProcessBufferProcessor] Unable to process message <fb393705-472e-11ea-bb88-4256e21dfc8e>: java.lang.ClassCastException: Cannot cast java.lang.String to org.joda.time.DateTime 2020-02-04T10:16:09.713+01:00 WARN [ProcessBufferProcessor] Unable to process message <fb393706-472e-11ea-bb88-4256e21dfc8e>: java.lang.ClassCastException: Cannot cast java.lang.String to org.joda.time.DateTime 2020-02-04T10:16:09.958+01:00 WARN [ProcessBufferProcessor] Unable to process message <fb5e7245-472e-11ea-bb88-4256e21dfc8e>: java.lang.ClassCastException: Cannot cast java.lang.String to org.joda.time.DateTime 2020-02-04T10:16:09.958+01:00 WARN [ProcessBufferProcessor] Unable to process message <fb5e7244-472e-11ea-bb88-4256e21dfc8e>: java.lang.ClassCastException: Cannot cast java.lang.String to org.joda.time.DateTime 2020-02-04T10:16:09.958+01:00 WARN [ProcessBufferProcessor] Unable to process message <fb5e7243-472e-11ea-bb88-4256e21dfc8e>: java.lang.ClassCastException: Cannot cast java.lang.String to org.joda.time.DateTime 2020-02-04T10:16:10.076+01:00 WARN [ProcessBufferProcessor] Unable to process message <fb709ab4-472e-11ea-bb88-4256e21dfc8e>: java.lang.ClassCastException: Cannot cast java.lang.String to org.joda.time.DateTime 2020-02-04T10:16:10.077+01:00 WARN [ProcessBufferProcessor] Unable to process message <fb70c1c0-472e-11ea-bb88-4256e21dfc8e>: java.lang.ClassCastException: Cannot cast java.lang.String to org.joda.time.DateTime ``` When this happened, I have lost a lot of logs that were not indexed in graylog. When I delete the extractor everything work smoothly. ## Possible Solution Catch the error so that the logs can be indexed even tho the extractor failed ? ## Steps to Reproduce (for bugs) 1. Graylog in 3.2 2. Add an extractor with the Grok Pattern from above 3. Add logs that do not match the Grok Pattern to Graylog ## Context This bug automatically dropped logs without warning. ## Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> * Graylog Version: 3.2 * Elasticsearch Version: * MongoDB Version: * Operating System: Linux/Debian * Browser version: Firefox
process
grok pattern extractors crashes and drop logs after i upgraded graylog to i found that one extractor containing a grok pattern was constantly crashing the logs were then forever lost expected behavior when i add an extractor with a grok pattern for syslog and select always try to extract graylog should not drop the logs that failed current behavior i added the following extractor and i can see in graylog server logs warn unable to process message java lang classcastexception cannot cast java lang string to org joda time datetime warn unable to process message java lang classcastexception cannot cast java lang string to org joda time datetime warn unable to process message java lang classcastexception cannot cast java lang string to org joda time datetime warn unable to process message java lang classcastexception cannot cast java lang string to org joda time datetime warn unable to process message java lang classcastexception cannot cast java lang string to org joda time datetime warn unable to process message java lang classcastexception cannot cast java lang string to org joda time datetime warn unable to process message java lang classcastexception cannot cast java lang string to org joda time datetime warn unable to process message java lang classcastexception cannot cast java lang string to org joda time datetime when this happened i have lost a lot of logs that were not indexed in graylog when i delete the extractor everything work smoothly possible solution catch the error so that the logs can be indexed even tho the extractor failed steps to reproduce for bugs graylog in add an extractor with the grok pattern from above add logs that do not match the grok pattern to graylog context this bug automatically dropped logs without warning your environment graylog version elasticsearch version mongodb version operating system linux debian browser version firefox
1
8,542
11,714,422,489
IssuesEvent
2020-03-09 12:19:13
nodejs/node
https://api.github.com/repos/nodejs/node
closed
docs: needs clarification: "terminal raw mode" (signal handling)
doc process tty
About the `SIGINT` event on `process` it is said in the docs that > It is not generated when terminal raw mode is enabled. https://github.com/nodejs/node/blame/3ec4b21b1c438255df6f1652377011080dc28052/doc/api/process.md#L504 In the example program I play around with the SIGINT handler was not firing. I assumed that I was testing with the "terminal raw mode", so I was looking into understanding what that is and how to disable it. A web search for `nodejs "terminal raw mode"` didn't yield anything useful, though. I also explored `node --help` and didn't see anything obvious. I think we should clarify in docs what "terminal raw mode" is and then cross-link to that place from the SIGINT doc I linked above.
1.0
docs: needs clarification: "terminal raw mode" (signal handling) - About the `SIGINT` event on `process` it is said in the docs that > It is not generated when terminal raw mode is enabled. https://github.com/nodejs/node/blame/3ec4b21b1c438255df6f1652377011080dc28052/doc/api/process.md#L504 In the example program I play around with the SIGINT handler was not firing. I assumed that I was testing with the "terminal raw mode", so I was looking into understanding what that is and how to disable it. A web search for `nodejs "terminal raw mode"` didn't yield anything useful, though. I also explored `node --help` and didn't see anything obvious. I think we should clarify in docs what "terminal raw mode" is and then cross-link to that place from the SIGINT doc I linked above.
process
docs needs clarification terminal raw mode signal handling about the sigint event on process it is said in the docs that it is not generated when terminal raw mode is enabled in the example program i play around with the sigint handler was not firing i assumed that i was testing with the terminal raw mode so i was looking into understanding what that is and how to disable it a web search for nodejs terminal raw mode didn t yield anything useful though i also explored node help and didn t see anything obvious i think we should clarify in docs what terminal raw mode is and then cross link to that place from the sigint doc i linked above
1
2,634
5,412,258,728
IssuesEvent
2017-03-01 14:07:31
neuropoly/spinalcordtoolbox
https://api.github.com/repos/neuropoly/spinalcordtoolbox
opened
NURBS are unstable when calculated in voxel coordinates
bug fix:minor priority: high sct_process_segmentation
especially when voxel spacing in one direction is very large. This instability causes fluctuations of the derivatives, which induces wrong calculations of the angles. This has an strong effect on CSA computation, as demonstrated in the example below. Data: ` ` Command: ``` sct_process_segmentation -i t2s_segm.nii.gz -p csa ``` Results: ``` # Slice (z),CSA (mm^2),Angle with respect to the I-S direction (degrees) 0,63.5923927147,18.6837707469 1,79.2658753014,8.26795055912 2,88.2874975915,5.12747654869 3,87.9995780207,6.03282331933 4,-8.61862123509,95.2875209491 5,89.2278470889,4.91847828944 6,95.8647986243,9.45846253778 7,84.0700371176,4.83530051485 8,84.7877521872,1.7651865876 9,81.1628827015,0.530803473898 10,82.0784541216,0.515470049064 11,80.383388821,1.2822922865 12,83.1066671944,1.84459163675 ```
1.0
NURBS are unstable when calculated in voxel coordinates - especially when voxel spacing in one direction is very large. This instability causes fluctuations of the derivatives, which induces wrong calculations of the angles. This has an strong effect on CSA computation, as demonstrated in the example below. Data: ` ` Command: ``` sct_process_segmentation -i t2s_segm.nii.gz -p csa ``` Results: ``` # Slice (z),CSA (mm^2),Angle with respect to the I-S direction (degrees) 0,63.5923927147,18.6837707469 1,79.2658753014,8.26795055912 2,88.2874975915,5.12747654869 3,87.9995780207,6.03282331933 4,-8.61862123509,95.2875209491 5,89.2278470889,4.91847828944 6,95.8647986243,9.45846253778 7,84.0700371176,4.83530051485 8,84.7877521872,1.7651865876 9,81.1628827015,0.530803473898 10,82.0784541216,0.515470049064 11,80.383388821,1.2822922865 12,83.1066671944,1.84459163675 ```
process
nurbs are unstable when calculated in voxel coordinates especially when voxel spacing in one direction is very large this instability causes fluctuations of the derivatives which induces wrong calculations of the angles this has an strong effect on csa computation as demonstrated in the example below data command sct process segmentation i segm nii gz p csa results slice z csa mm angle with respect to the i s direction degrees
1
14,811
18,143,492,276
IssuesEvent
2021-09-25 02:39:25
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Crash when clipping a larger layer with a small layer
Feedback stale Processing Bug Crash/Data Corruption Upstream
### What is the bug or the crash? Qgis crashes after approx 10s when clipping a larger layer with a small overlay layer. Restarting Qgis or rebooting the computer doesn't help. ![grafik](https://user-images.githubusercontent.com/67278094/128853021-1ea4ad4d-3d58-4dec-8061-49c0e5c3c26c.png) Input layer (Feature count: 2.923.688, CRS ESRI:54009 - World_Mollweide): [GHS_SMOD_Polygonized_raster2vector.zip](https://drive.google.com/file/d/1QJqkm585TIgtG7J8skGN3nLj7E7SLELO/view?usp=sharing) Overlay layer (Feature count: 1, CRS EPSG:4326 - WGS 84 ): [smod_polygonized_raster2vector_smallMaskLayer.zip](https://github.com/qgis/QGIS/files/6960914/smod_polygonized_raster2vector_smallMaskLayer.zip) ### Steps to reproduce the issue 1. Import both layers into a new qgis project 2. processing tools -> clip 3. choose layers according to the screenshot 4. a crash occurs after several seconds ### Versions QGIS version | 3.20.1-Odense | QGIS code revision | 1c3c5cd6 -- | -- | -- | -- Qt version | 5.15.2 Python version | 3.9.5 GDAL/OGR version | 3.3.1 PROJ version | 8.1.0 EPSG Registry database version | v10.027 (2021-06-17) GEOS version | 3.9.1-CAPI-1.14.2 SQLite version | 3.35.2 PDAL version | 2.3.0 PostgreSQL client version | 13.0 SpatiaLite version | 5.0.1 QWT version | 6.1.3 QScintilla2 version | 2.11.5 OS version | Windows 10 Version 2009   |   |   |   Active Python plugins | GroupStatslatlontoolsnominatimQuickOSMdb_managerprocessing ### Additional context _No response_
1.0
Crash when clipping a larger layer with a small layer - ### What is the bug or the crash? Qgis crashes after approx 10s when clipping a larger layer with a small overlay layer. Restarting Qgis or rebooting the computer doesn't help. ![grafik](https://user-images.githubusercontent.com/67278094/128853021-1ea4ad4d-3d58-4dec-8061-49c0e5c3c26c.png) Input layer (Feature count: 2.923.688, CRS ESRI:54009 - World_Mollweide): [GHS_SMOD_Polygonized_raster2vector.zip](https://drive.google.com/file/d/1QJqkm585TIgtG7J8skGN3nLj7E7SLELO/view?usp=sharing) Overlay layer (Feature count: 1, CRS EPSG:4326 - WGS 84 ): [smod_polygonized_raster2vector_smallMaskLayer.zip](https://github.com/qgis/QGIS/files/6960914/smod_polygonized_raster2vector_smallMaskLayer.zip) ### Steps to reproduce the issue 1. Import both layers into a new qgis project 2. processing tools -> clip 3. choose layers according to the screenshot 4. a crash occurs after several seconds ### Versions QGIS version | 3.20.1-Odense | QGIS code revision | 1c3c5cd6 -- | -- | -- | -- Qt version | 5.15.2 Python version | 3.9.5 GDAL/OGR version | 3.3.1 PROJ version | 8.1.0 EPSG Registry database version | v10.027 (2021-06-17) GEOS version | 3.9.1-CAPI-1.14.2 SQLite version | 3.35.2 PDAL version | 2.3.0 PostgreSQL client version | 13.0 SpatiaLite version | 5.0.1 QWT version | 6.1.3 QScintilla2 version | 2.11.5 OS version | Windows 10 Version 2009   |   |   |   Active Python plugins | GroupStatslatlontoolsnominatimQuickOSMdb_managerprocessing ### Additional context _No response_
process
crash when clipping a larger layer with a small layer what is the bug or the crash qgis crashes after approx when clipping a larger layer with a small overlay layer restarting qgis or rebooting the computer doesn t help input layer feature count crs esri world mollweide overlay layer feature count crs epsg wgs steps to reproduce the issue import both layers into a new qgis project processing tools clip choose layers according to the screenshot a crash occurs after several seconds versions qgis version odense qgis code revision qt version python version gdal ogr version proj version epsg registry database version geos version capi sqlite version pdal version postgresql client version spatialite version qwt version version os version windows version         active python plugins groupstatslatlontoolsnominatimquickosmdb managerprocessing additional context no response
1
100,908
30,813,173,077
IssuesEvent
2023-08-01 11:51:38
assistant-ai/jess
https://api.github.com/repos/assistant-ai/jess
reopened
Cloudbuild: use Windows container for `go build` with `CGO_ENABLED=1`
feature build
Maybe it is the best container. ) https://hub.docker.com/layers/library/golang/windowsservercore-ltsc2022/images/sha256-647b841b8cc8b449ebd00e2774b7fcc8753d7053dd83227c11c306d956662f00?context=explore
1.0
Cloudbuild: use Windows container for `go build` with `CGO_ENABLED=1` - Maybe it is the best container. ) https://hub.docker.com/layers/library/golang/windowsservercore-ltsc2022/images/sha256-647b841b8cc8b449ebd00e2774b7fcc8753d7053dd83227c11c306d956662f00?context=explore
non_process
cloudbuild use windows container for go build with cgo enabled maybe it is the best container
0
303,720
23,037,141,865
IssuesEvent
2022-07-22 20:14:23
stoplightio/prism
https://api.github.com/repos/stoplightio/prism
closed
Add protocol requirement to prism proxy URLs
documentation
This topic: https://meta.stoplight.io/docs/prism/72d69fb629de0-validation-proxy#validation-proxy Has an incorrect example. prism proxy reference/backend/openapi.yaml localhost:3000 --errors Should be: prism proxy reference/backend/openapi.yaml http://localhost:3000 --errors
1.0
Add protocol requirement to prism proxy URLs - This topic: https://meta.stoplight.io/docs/prism/72d69fb629de0-validation-proxy#validation-proxy Has an incorrect example. prism proxy reference/backend/openapi.yaml localhost:3000 --errors Should be: prism proxy reference/backend/openapi.yaml http://localhost:3000 --errors
non_process
add protocol requirement to prism proxy urls this topic has an incorrect example prism proxy reference backend openapi yaml localhost errors should be prism proxy reference backend openapi yaml errors
0
16,511
21,519,625,306
IssuesEvent
2022-04-28 13:10:55
symfony/symfony
https://api.github.com/repos/symfony/symfony
closed
[Process] Allow running multiple commands at once
Feature Process Stalled
**Description** When running a lot of small shell commands after each other, the overhead of creating the underlying processes becomes noticeable. It would be handy if the process component would support chaining commands. Working directly on a Linux/Windows shell you would typically do this: ```shell # Linux $ foo; bar; baz $ foo && bar && baz # Windows $ foo & bar & baz $ foo && bar && baz ``` (Yes there are other chaining operators as well like `||`.) This is currently not possible with the process component. **API Example** I like that `Symfony\Component\Process\Process` became somewhat immutable now (the command line is baked in after the constructor ran), so the API could maybe look like this: ```php $process = new Process(['foo']); // foo; bar $process = $process->withChainedCommand(['bar']); // returns a new process instance // foo && bar $process = $process->withAndChainedCommand(['bar']); // returns a new process instance ``` What do you think? Is this a worthwhile addition?
1.0
[Process] Allow running multiple commands at once - **Description** When running a lot of small shell commands after each other, the overhead of creating the underlying processes becomes noticeable. It would be handy if the process component would support chaining commands. Working directly on a Linux/Windows shell you would typically do this: ```shell # Linux $ foo; bar; baz $ foo && bar && baz # Windows $ foo & bar & baz $ foo && bar && baz ``` (Yes there are other chaining operators as well like `||`.) This is currently not possible with the process component. **API Example** I like that `Symfony\Component\Process\Process` became somewhat immutable now (the command line is baked in after the constructor ran), so the API could maybe look like this: ```php $process = new Process(['foo']); // foo; bar $process = $process->withChainedCommand(['bar']); // returns a new process instance // foo && bar $process = $process->withAndChainedCommand(['bar']); // returns a new process instance ``` What do you think? Is this a worthwhile addition?
process
allow running multiple commands at once description when running a lot of small shell commands after each other the overhead of creating the underlying processes becomes noticeable it would be handy if the process component would support chaining commands working directly on a linux windows shell you would typically do this shell linux foo bar baz foo bar baz windows foo bar baz foo bar baz yes there are other chaining operators as well like this is currently not possible with the process component api example i like that symfony component process process became somewhat immutable now the command line is baked in after the constructor ran so the api could maybe look like this php process new process foo bar process process withchainedcommand returns a new process instance foo bar process process withandchainedcommand returns a new process instance what do you think is this a worthwhile addition
1
67,592
27,958,769,316
IssuesEvent
2023-03-24 14:14:59
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Chat Completion API
cognitive-services/svc triaged cxp doc-enhancement Pri1
Hello, I would like to know if "Chat Completion API" means "an API for [chat completion](https://platform.openai.com/docs/guides/chat)" (not a proper noun), or if it's the name of an API, which will remain in English, even for the other languages (which means this "Chat Completion API" name should not be translated). Thanks! Best regards, François --- #### Document Details ⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.* * ID: 13994f3e-a96a-8804-4deb-7e0adad7f479 * Version Independent ID: e087d8ce-bf3a-f12c-41dc-3ea924c6f089 * Content: [How to work with the ChatGPT and GPT-4 models (preview) - Azure OpenAI Service](https://learn.microsoft.com/en-us/azure/cognitive-services/openai/how-to/chatgpt?pivots=programming-language-chat-completions#working-with-the-chat-completion-api) * Content Source: [articles/cognitive-services/openai/how-to/chatgpt.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/cognitive-services/openai/how-to/chatgpt.md) * Service: **cognitive-services** * GitHub Login: @mrbullwinkle * Microsoft Alias: **mbullwin**
1.0
Chat Completion API - Hello, I would like to know if "Chat Completion API" means "an API for [chat completion](https://platform.openai.com/docs/guides/chat)" (not a proper noun), or if it's the name of an API, which will remain in English, even for the other languages (which means this "Chat Completion API" name should not be translated). Thanks! Best regards, François --- #### Document Details ⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.* * ID: 13994f3e-a96a-8804-4deb-7e0adad7f479 * Version Independent ID: e087d8ce-bf3a-f12c-41dc-3ea924c6f089 * Content: [How to work with the ChatGPT and GPT-4 models (preview) - Azure OpenAI Service](https://learn.microsoft.com/en-us/azure/cognitive-services/openai/how-to/chatgpt?pivots=programming-language-chat-completions#working-with-the-chat-completion-api) * Content Source: [articles/cognitive-services/openai/how-to/chatgpt.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/cognitive-services/openai/how-to/chatgpt.md) * Service: **cognitive-services** * GitHub Login: @mrbullwinkle * Microsoft Alias: **mbullwin**
non_process
chat completion api hello i would like to know if chat completion api means an api for not a proper noun or if it s the name of an api which will remain in english even for the other languages which means this chat completion api name should not be translated thanks best regards françois document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id version independent id content content source service cognitive services github login mrbullwinkle microsoft alias mbullwin
0
22,198
30,755,628,325
IssuesEvent
2023-07-29 02:40:47
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
[processor/resourcedetection] No system attributes set if `host.id` cannot be fetched
bug priority:p1 processor/resourcedetection
### Component(s) _No response_ ### What happened? After https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/24239, instead of setting empty `host.id`, we started throwing an error and dropping all other important system detector attributes: `host.name`, `os.type`, `host.arch`. We need set other system resources even if `host.id` cannot be fetched. Also, it's happening even if the `host.id` is disabled, which is now the default behavior. https://github.com/open-telemetry/opentelemetry-collector-contrib/issues/24369 will solve this particular issue. But it's not as important as this one. ### Collector version 0.82.0 (binaries are not released yet)
1.0
[processor/resourcedetection] No system attributes set if `host.id` cannot be fetched - ### Component(s) _No response_ ### What happened? After https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/24239, instead of setting empty `host.id`, we started throwing an error and dropping all other important system detector attributes: `host.name`, `os.type`, `host.arch`. We need set other system resources even if `host.id` cannot be fetched. Also, it's happening even if the `host.id` is disabled, which is now the default behavior. https://github.com/open-telemetry/opentelemetry-collector-contrib/issues/24369 will solve this particular issue. But it's not as important as this one. ### Collector version 0.82.0 (binaries are not released yet)
process
no system attributes set if host id cannot be fetched component s no response what happened after instead of setting empty host id we started throwing an error and dropping all other important system detector attributes host name os type host arch we need set other system resources even if host id cannot be fetched also it s happening even if the host id is disabled which is now the default behavior will solve this particular issue but it s not as important as this one collector version binaries are not released yet
1
12,209
14,742,829,363
IssuesEvent
2021-01-07 12:58:01
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
68 Portland - SAB Latency
anc-process anp-2 ant-support
In GitLab by @kdjstudios on Jun 7, 2019, 09:22 **Submitted by:** "Grant Crymes" <grant.crymes@answernet.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/8513066 - CLOSED **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/10208596 **Server:** Internal **Client/Site:** Portland **Account:** NA **Issue:** We are experiencing latency when posting checks in SAB. It’s taking approximately 8-20 seconds to advance through each screen which is really slowing the payment processing down. This latency issue started a month or so ago but has been getting progressively worse.
1.0
68 Portland - SAB Latency - In GitLab by @kdjstudios on Jun 7, 2019, 09:22 **Submitted by:** "Grant Crymes" <grant.crymes@answernet.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/8513066 - CLOSED **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/10208596 **Server:** Internal **Client/Site:** Portland **Account:** NA **Issue:** We are experiencing latency when posting checks in SAB. It’s taking approximately 8-20 seconds to advance through each screen which is really slowing the payment processing down. This latency issue started a month or so ago but has been getting progressively worse.
process
portland sab latency in gitlab by kdjstudios on jun submitted by grant crymes helpdesk closed helpdesk server internal client site portland account na issue we are experiencing latency when posting checks in sab it’s taking approximately seconds to advance through each screen which is really slowing the payment processing down this latency issue started a month or so ago but has been getting progressively worse
1
22,089
7,113,517,689
IssuesEvent
2018-01-17 20:46:12
dotnet/buildtools
https://api.github.com/repos/dotnet/buildtools
closed
PushToBlobFeed manifest creation should use VSTS default variables from env as a fallback
area-building-support
For Core repos (and maybe more) these values don't match what we need, but adding this fallback would make the implementation in some places have less code. (See https://github.com/dotnet/core-eng/issues/2404#issuecomment-358389341.) https://docs.microsoft.com/en-us/vsts/build-release/concepts/definitions/build/variables * `BUILD_REPOSITORY_NAME` * `BUILD_SOURCEVERSION` * `BUILD_SOURCEBRANCHNAME` * In my experience this has chopped off the `release` part of `release/2.0` so it might not work. `BUILD_SOURCEBRANCH` contains more info than needed but might be better. * (Edit) Confirmed that in a CoreCLR `release/2.0.0` build, this var is `2.0.0`. SOURCEBRANCH is `refs/heads/release/2.0.0`. * `BUILD_BUILDID` /cc @tmat @tannergooding
1.0
PushToBlobFeed manifest creation should use VSTS default variables from env as a fallback - For Core repos (and maybe more) these values don't match what we need, but adding this fallback would make the implementation in some places have less code. (See https://github.com/dotnet/core-eng/issues/2404#issuecomment-358389341.) https://docs.microsoft.com/en-us/vsts/build-release/concepts/definitions/build/variables * `BUILD_REPOSITORY_NAME` * `BUILD_SOURCEVERSION` * `BUILD_SOURCEBRANCHNAME` * In my experience this has chopped off the `release` part of `release/2.0` so it might not work. `BUILD_SOURCEBRANCH` contains more info than needed but might be better. * (Edit) Confirmed that in a CoreCLR `release/2.0.0` build, this var is `2.0.0`. SOURCEBRANCH is `refs/heads/release/2.0.0`. * `BUILD_BUILDID` /cc @tmat @tannergooding
non_process
pushtoblobfeed manifest creation should use vsts default variables from env as a fallback for core repos and maybe more these values don t match what we need but adding this fallback would make the implementation in some places have less code see build repository name build sourceversion build sourcebranchname in my experience this has chopped off the release part of release so it might not work build sourcebranch contains more info than needed but might be better edit confirmed that in a coreclr release build this var is sourcebranch is refs heads release build buildid cc tmat tannergooding
0
170,768
14,269,525,516
IssuesEvent
2020-11-21 01:51:17
AzureAD/microsoft-authentication-library-for-js
https://api.github.com/repos/AzureAD/microsoft-authentication-library-for-js
opened
Modified copy of msal-angular code inside samples/msal-angular-v2-samples/angular10-browser-sample?
documentation question
## Library - [x] `@azure/msal-browser@2.x.x` - [x] `@azure/msal-angular@2.x.x` ## Documentation location - [x] Documentation does not exist ## Description The Angular 10 example does not use the code from `@azure/msal-angular@2.x.x`. Instead it has copied and modified the code, see https://github.com/AzureAD/microsoft-authentication-library-for-js/tree/dev/samples/msal-angular-v2-samples/angular10-browser-sample/src/app/msal. Why is that? Does it mean we can't use the official `@azure/msal-angular@2.x.x` library, but instead have to copy the code from the example? What happens with bug fixes? Will they be committed to both, @azure/msal-angular@2.x.x` and the example project?
1.0
Modified copy of msal-angular code inside samples/msal-angular-v2-samples/angular10-browser-sample? - ## Library - [x] `@azure/msal-browser@2.x.x` - [x] `@azure/msal-angular@2.x.x` ## Documentation location - [x] Documentation does not exist ## Description The Angular 10 example does not use the code from `@azure/msal-angular@2.x.x`. Instead it has copied and modified the code, see https://github.com/AzureAD/microsoft-authentication-library-for-js/tree/dev/samples/msal-angular-v2-samples/angular10-browser-sample/src/app/msal. Why is that? Does it mean we can't use the official `@azure/msal-angular@2.x.x` library, but instead have to copy the code from the example? What happens with bug fixes? Will they be committed to both, @azure/msal-angular@2.x.x` and the example project?
non_process
modified copy of msal angular code inside samples msal angular samples browser sample library azure msal browser x x azure msal angular x x documentation location documentation does not exist description the angular example does not use the code from azure msal angular x x instead it has copied and modified the code see why is that does it mean we can t use the official azure msal angular x x library but instead have to copy the code from the example what happens with bug fixes will they be committed to both azure msal angular x x and the example project
0
22,552
31,761,984,305
IssuesEvent
2023-09-12 06:09:47
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
[Mirror] rules_graalvm@v0.10.2
P2 type: process team-OSS mirror request
### Please list the URLs of the archives you'd like to mirror: https://github.com/sgammon/rules_graalvm/releases/download/v0.10.2/rules_graalvm-0.10.2.zip https://github.com/graalvm/graalvm-ce-builds/releases/download/graal-23.0.1/native-image-installable-svm-java20-windows-amd64-23.0.1.jar https://github.com/graalvm/graalvm-ce-builds/releases/download/jdk-20.0.2/graalvm-community-jdk-20.0.2_windows-x64_bin.zip https://github.com/graalvm/graalvm-ce-builds/releases/download/graal-23.0.1/native-image-installable-svm-java20-linux-amd64-23.0.1.jar https://github.com/graalvm/graalvm-ce-builds/releases/download/jdk-20.0.2/graalvm-community-jdk-20.0.2_linux-x64_bin.tar.gz https://github.com/graalvm/graalvm-ce-builds/releases/download/graal-23.0.1/native-image-installable-svm-java20-darwin-aarch64-23.0.1.jar https://github.com/graalvm/graalvm-ce-builds/releases/download/jdk-20.0.2/graalvm-community-jdk-20.0.2_macos-aarch64_bin.tar.gz https://github.com/graalvm/graalvm-ce-builds/releases/download/graal-23.0.1/native-image-installable-svm-java20-darwin-amd64-23.0.1.jar https://github.com/graalvm/graalvm-ce-builds/releases/download/jdk-20.0.2/graalvm-community-jdk-20.0.2_macos-x64_bin.tar.gz
1.0
[Mirror] rules_graalvm@v0.10.2 - ### Please list the URLs of the archives you'd like to mirror: https://github.com/sgammon/rules_graalvm/releases/download/v0.10.2/rules_graalvm-0.10.2.zip https://github.com/graalvm/graalvm-ce-builds/releases/download/graal-23.0.1/native-image-installable-svm-java20-windows-amd64-23.0.1.jar https://github.com/graalvm/graalvm-ce-builds/releases/download/jdk-20.0.2/graalvm-community-jdk-20.0.2_windows-x64_bin.zip https://github.com/graalvm/graalvm-ce-builds/releases/download/graal-23.0.1/native-image-installable-svm-java20-linux-amd64-23.0.1.jar https://github.com/graalvm/graalvm-ce-builds/releases/download/jdk-20.0.2/graalvm-community-jdk-20.0.2_linux-x64_bin.tar.gz https://github.com/graalvm/graalvm-ce-builds/releases/download/graal-23.0.1/native-image-installable-svm-java20-darwin-aarch64-23.0.1.jar https://github.com/graalvm/graalvm-ce-builds/releases/download/jdk-20.0.2/graalvm-community-jdk-20.0.2_macos-aarch64_bin.tar.gz https://github.com/graalvm/graalvm-ce-builds/releases/download/graal-23.0.1/native-image-installable-svm-java20-darwin-amd64-23.0.1.jar https://github.com/graalvm/graalvm-ce-builds/releases/download/jdk-20.0.2/graalvm-community-jdk-20.0.2_macos-x64_bin.tar.gz
process
rules graalvm please list the urls of the archives you d like to mirror
1
2,878
5,833,394,325
IssuesEvent
2017-05-09 01:24:04
codefordenver/org
https://api.github.com/repos/codefordenver/org
closed
Create Mission Statement and upload to all appropriate places
Process
Create mission statement to aid fundraising marketing campaign and general understanding of organization. Get sign-off from heads.
1.0
Create Mission Statement and upload to all appropriate places - Create mission statement to aid fundraising marketing campaign and general understanding of organization. Get sign-off from heads.
process
create mission statement and upload to all appropriate places create mission statement to aid fundraising marketing campaign and general understanding of organization get sign off from heads
1
133,771
18,353,242,454
IssuesEvent
2021-10-08 14:51:52
symfony/symfony-docs
https://api.github.com/repos/symfony/symfony-docs
closed
[Security] Automatically register custom authenticator as entry_point (…
Security
| Q | A | ------------ | --- | Feature PR | symfony/symfony#39153 | PR author(s) | @wouterj | Merged in | 5.2
True
[Security] Automatically register custom authenticator as entry_point (… - | Q | A | ------------ | --- | Feature PR | symfony/symfony#39153 | PR author(s) | @wouterj | Merged in | 5.2
non_process
automatically register custom authenticator as entry point … q a feature pr symfony symfony pr author s wouterj merged in
0
216,706
7,311,092,383
IssuesEvent
2018-02-28 16:44:53
EthereumCommonwealth/etherwallet
https://api.github.com/repos/EthereumCommonwealth/etherwallet
closed
Add more networks.
low_priority
### Description Just add custom nodes from https://github.com/kvhnuke/etherwallet/blob/mercury/app/scripts/nodes.js - Tomo Coin network. - Ella network. - POA network. NOTE: It may be necessary to configure corresponding colorings for each of the networks.
1.0
Add more networks. - ### Description Just add custom nodes from https://github.com/kvhnuke/etherwallet/blob/mercury/app/scripts/nodes.js - Tomo Coin network. - Ella network. - POA network. NOTE: It may be necessary to configure corresponding colorings for each of the networks.
non_process
add more networks description just add custom nodes from tomo coin network ella network poa network note it may be necessary to configure corresponding colorings for each of the networks
0
178,030
29,486,349,611
IssuesEvent
2023-06-02 10:01:28
OfficeDev/TeamsFx
https://api.github.com/repos/OfficeDev/TeamsFx
closed
TeamsFx spfx tab project Env's replicating each other
needs more info *as-designed TA:E2E no recent activity
**Describe the bug** Hi, I have created a spfx tab project from teamfx cli. i have created two env's dev & test which i am managing from teams toolkit exstension. when i deploy new update in dev env and approve it at MS teams admin portal it's automatically reflecting me at both dev & test env app. Also, can't use .env file in spfx tab project. **To Reproduce** Steps to reproduce the behavior: 1. Create new spfx tab project from teamsfx cli. 2. Create new env's and provision them. 3. Deploy both env's app and approve at MS teams admin 4. Create new update for dev env and deploy it. 5. It reflects at both dev & test env apps. **Expected behavior** Dev update should not be reflect at other env if we deploy new update only at dev env. **VS Code Extension Information (please complete the following information):** - Teams Toolkit extension verion: 4.1.3 **CLI Information (please complete the following information):** - Teamsfx-cli version: 1.2.4
1.0
TeamsFx spfx tab project Env's replicating each other - **Describe the bug** Hi, I have created a spfx tab project from teamfx cli. i have created two env's dev & test which i am managing from teams toolkit exstension. when i deploy new update in dev env and approve it at MS teams admin portal it's automatically reflecting me at both dev & test env app. Also, can't use .env file in spfx tab project. **To Reproduce** Steps to reproduce the behavior: 1. Create new spfx tab project from teamsfx cli. 2. Create new env's and provision them. 3. Deploy both env's app and approve at MS teams admin 4. Create new update for dev env and deploy it. 5. It reflects at both dev & test env apps. **Expected behavior** Dev update should not be reflect at other env if we deploy new update only at dev env. **VS Code Extension Information (please complete the following information):** - Teams Toolkit extension verion: 4.1.3 **CLI Information (please complete the following information):** - Teamsfx-cli version: 1.2.4
non_process
teamsfx spfx tab project env s replicating each other describe the bug hi i have created a spfx tab project from teamfx cli i have created two env s dev test which i am managing from teams toolkit exstension when i deploy new update in dev env and approve it at ms teams admin portal it s automatically reflecting me at both dev test env app also can t use env file in spfx tab project to reproduce steps to reproduce the behavior create new spfx tab project from teamsfx cli create new env s and provision them deploy both env s app and approve at ms teams admin create new update for dev env and deploy it it reflects at both dev test env apps expected behavior dev update should not be reflect at other env if we deploy new update only at dev env vs code extension information please complete the following information teams toolkit extension verion cli information please complete the following information teamsfx cli version
0
50,256
13,187,403,496
IssuesEvent
2020-08-13 03:18:23
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
closed
make_tarball_rootsys.sh.in bugfix (Trac #393)
Migrated from Trac combo core defect
cmake/make_tarball_rootsys.sh.in needs a recursive copy flag '-r' since new ROOT lib dir contains subdirectories. Line 14: - cp -P @ROOTSYS@/lib/* @CMAKE_INSTALL_PREFIX@/cernroot/lib + cp -r -P @ROOTSYS@/lib/* @CMAKE_INSTALL_PREFIX@/cernroot/lib <details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/393 , reported by juancarlos and owned by blaufuss_</summary> <p> ```json { "status": "closed", "changetime": "2012-06-03T04:44:16", "description": "cmake/make_tarball_rootsys.sh.in needs a recursive copy flag '-r' since new ROOT lib dir contains subdirectories.\n\nLine 14:\n\n- cp -P @ROOTSYS@/lib/* @CMAKE_INSTALL_PREFIX@/cernroot/lib\n+ cp -r -P @ROOTSYS@/lib/* @CMAKE_INSTALL_PREFIX@/cernroot/lib", "reporter": "juancarlos", "cc": "", "resolution": "fixed", "_ts": "1338698656000000", "component": "combo core", "summary": "make_tarball_rootsys.sh.in bugfix", "priority": "normal", "keywords": "cmake", "time": "2012-05-16T21:21:39", "milestone": "", "owner": "blaufuss", "type": "defect" } ``` </p> </details>
1.0
make_tarball_rootsys.sh.in bugfix (Trac #393) - cmake/make_tarball_rootsys.sh.in needs a recursive copy flag '-r' since new ROOT lib dir contains subdirectories. Line 14: - cp -P @ROOTSYS@/lib/* @CMAKE_INSTALL_PREFIX@/cernroot/lib + cp -r -P @ROOTSYS@/lib/* @CMAKE_INSTALL_PREFIX@/cernroot/lib <details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/393 , reported by juancarlos and owned by blaufuss_</summary> <p> ```json { "status": "closed", "changetime": "2012-06-03T04:44:16", "description": "cmake/make_tarball_rootsys.sh.in needs a recursive copy flag '-r' since new ROOT lib dir contains subdirectories.\n\nLine 14:\n\n- cp -P @ROOTSYS@/lib/* @CMAKE_INSTALL_PREFIX@/cernroot/lib\n+ cp -r -P @ROOTSYS@/lib/* @CMAKE_INSTALL_PREFIX@/cernroot/lib", "reporter": "juancarlos", "cc": "", "resolution": "fixed", "_ts": "1338698656000000", "component": "combo core", "summary": "make_tarball_rootsys.sh.in bugfix", "priority": "normal", "keywords": "cmake", "time": "2012-05-16T21:21:39", "milestone": "", "owner": "blaufuss", "type": "defect" } ``` </p> </details>
non_process
make tarball rootsys sh in bugfix trac cmake make tarball rootsys sh in needs a recursive copy flag r since new root lib dir contains subdirectories line cp p rootsys lib cmake install prefix cernroot lib cp r p rootsys lib cmake install prefix cernroot lib migrated from reported by juancarlos and owned by blaufuss json status closed changetime description cmake make tarball rootsys sh in needs a recursive copy flag r since new root lib dir contains subdirectories n nline n n cp p rootsys lib cmake install prefix cernroot lib n cp r p rootsys lib cmake install prefix cernroot lib reporter juancarlos cc resolution fixed ts component combo core summary make tarball rootsys sh in bugfix priority normal keywords cmake time milestone owner blaufuss type defect
0
19,874
26,288,037,800
IssuesEvent
2023-01-08 03:17:44
fdhhhdjd/Web-Online-School-Libary-Book
https://api.github.com/repos/fdhhhdjd/Web-Online-School-Libary-Book
opened
Read File Excel ( Back-End )
DEV Processing
**** Read File Excel **** - Library node-xlsx - Code performance equal buffer. - convert data JSON.
1.0
Read File Excel ( Back-End ) - **** Read File Excel **** - Library node-xlsx - Code performance equal buffer. - convert data JSON.
process
read file excel back end read file excel library node xlsx code performance equal buffer convert data json
1
21,524
29,806,279,230
IssuesEvent
2023-06-16 11:56:57
parca-dev/parca-agent
https://api.github.com/repos/parca-dev/parca-agent
closed
Normalization issues
P0 area/process-mapping
i.e. while running `kubectl run -n parca debug --image=python:latest -it` aka "the cockroachdb bug" (still present in `main`)
1.0
Normalization issues - i.e. while running `kubectl run -n parca debug --image=python:latest -it` aka "the cockroachdb bug" (still present in `main`)
process
normalization issues i e while running kubectl run n parca debug image python latest it aka the cockroachdb bug still present in main
1
296,626
22,310,482,100
IssuesEvent
2022-06-13 16:31:04
minetest/minetest
https://api.github.com/repos/minetest/minetest
closed
moveresult.touching_ground value is incorrect or misleading
@ Documentation Question
##### Minetest version <!-- Paste Minetest version between quotes below If you are on a devel version, please add git commit hash You can use `minetest --version` to find it. --> ``` Minetest 5.3.0 (Linux) Using Irrlicht 1.8.4 BUILD_TYPE=Release RUN_IN_PLACE=0 USE_GETTEXT=1 USE_SOUND=1 USE_CURL=1 USE_FREETYPE=1 USE_LUAJIT=1 STATIC_SHAREDIR="/nix/store/lyl5b6q7rvv7pp6szn0krddy7102fnx6-minetest-5.3.0/share/minetest" STATIC_LOCALEDIR="/nix/store/lyl5b6q7rvv7pp6szn0krddy7102fnx6-minetest-5.3.0/share/locale" ``` I've also tested with commit `4fcd000e20a26120349184cb9d40342b7876e6b8` from January 21st. ##### First of all Thanks for all the hard work. I'm loving writing mods for minetest. ##### Summary <!-- Describe your problem here --> The collision info table received during the `on_step` function has a wrong or misleading value for `touching_ground`. This value only seems to be `true` if the entity has just moved. For instance, you can spawn a physical entity in the air and set it's velocity to negative y. It will fall and `touching_ground` will be `false`, which is correct. Once it touches the ground, `touching_ground` becomes `true`, which is also correct, and the velocity seems to be reset to 0 because of the collision. The problem is that on the next `on_step`, `touching_ground` is now `false`, even though the entity is literally on the ground. My guess is that collisions were not calculated because there is no movement, but somehow the engine didn't account for the fact that the entity is *already* on the ground. This seems like a weird behavior to me, and I couldn't find a good way to determine that an entity is on the ground from Lua. Am I missing something here? If there's a good solution for this, I'm happy to write documentation about it.
1.0
moveresult.touching_ground value is incorrect or misleading - ##### Minetest version <!-- Paste Minetest version between quotes below If you are on a devel version, please add git commit hash You can use `minetest --version` to find it. --> ``` Minetest 5.3.0 (Linux) Using Irrlicht 1.8.4 BUILD_TYPE=Release RUN_IN_PLACE=0 USE_GETTEXT=1 USE_SOUND=1 USE_CURL=1 USE_FREETYPE=1 USE_LUAJIT=1 STATIC_SHAREDIR="/nix/store/lyl5b6q7rvv7pp6szn0krddy7102fnx6-minetest-5.3.0/share/minetest" STATIC_LOCALEDIR="/nix/store/lyl5b6q7rvv7pp6szn0krddy7102fnx6-minetest-5.3.0/share/locale" ``` I've also tested with commit `4fcd000e20a26120349184cb9d40342b7876e6b8` from January 21st. ##### First of all Thanks for all the hard work. I'm loving writing mods for minetest. ##### Summary <!-- Describe your problem here --> The collision info table received during the `on_step` function has a wrong or misleading value for `touching_ground`. This value only seems to be `true` if the entity has just moved. For instance, you can spawn a physical entity in the air and set it's velocity to negative y. It will fall and `touching_ground` will be `false`, which is correct. Once it touches the ground, `touching_ground` becomes `true`, which is also correct, and the velocity seems to be reset to 0 because of the collision. The problem is that on the next `on_step`, `touching_ground` is now `false`, even though the entity is literally on the ground. My guess is that collisions were not calculated because there is no movement, but somehow the engine didn't account for the fact that the entity is *already* on the ground. This seems like a weird behavior to me, and I couldn't find a good way to determine that an entity is on the ground from Lua. Am I missing something here? If there's a good solution for this, I'm happy to write documentation about it.
non_process
moveresult touching ground value is incorrect or misleading minetest version paste minetest version between quotes below if you are on a devel version please add git commit hash you can use minetest version to find it minetest linux using irrlicht build type release run in place use gettext use sound use curl use freetype use luajit static sharedir nix store minetest share minetest static localedir nix store minetest share locale i ve also tested with commit from january first of all thanks for all the hard work i m loving writing mods for minetest summary the collision info table received during the on step function has a wrong or misleading value for touching ground this value only seems to be true if the entity has just moved for instance you can spawn a physical entity in the air and set it s velocity to negative y it will fall and touching ground will be false which is correct once it touches the ground touching ground becomes true which is also correct and the velocity seems to be reset to because of the collision the problem is that on the next on step touching ground is now false even though the entity is literally on the ground my guess is that collisions were not calculated because there is no movement but somehow the engine didn t account for the fact that the entity is already on the ground this seems like a weird behavior to me and i couldn t find a good way to determine that an entity is on the ground from lua am i missing something here if there s a good solution for this i m happy to write documentation about it
0
6,508
6,490,278,376
IssuesEvent
2017-08-21 06:38:15
camptocamp/c2cgeoportal
https://api.github.com/repos/camptocamp/c2cgeoportal
closed
Suspicious relativ path to build the project standalone
Infrastructure Ready
Appears on c2cgeoportal master (20 feb 2017) I've tried to to a make build on the project but these two lines make issues: - https://github.com/camptocamp/c2cgeoportal/blob/daca42c1b3469531b7dc176ed3ffa5d368ede658/Makefile#L1 (`/build` shouldn't be `build` ?) - https://github.com/camptocamp/c2cgeoportal/blob/daca42c1b3469531b7dc176ed3ffa5d368ede658/Makefile#L61 ? Absolute path... that's not false ?
1.0
Suspicious relativ path to build the project standalone - Appears on c2cgeoportal master (20 feb 2017) I've tried to to a make build on the project but these two lines make issues: - https://github.com/camptocamp/c2cgeoportal/blob/daca42c1b3469531b7dc176ed3ffa5d368ede658/Makefile#L1 (`/build` shouldn't be `build` ?) - https://github.com/camptocamp/c2cgeoportal/blob/daca42c1b3469531b7dc176ed3ffa5d368ede658/Makefile#L61 ? Absolute path... that's not false ?
non_process
suspicious relativ path to build the project standalone appears on master feb i ve tried to to a make build on the project but these two lines make issues build shouldn t be build absolute path that s not false
0
411,031
27,811,102,130
IssuesEvent
2023-03-18 05:33:39
Real-Dev-Squad/website-api-contracts
https://api.github.com/repos/Real-Dev-Squad/website-api-contracts
closed
API Contract for GET Idle users/members
documentation
### AC - Create an API contract for `/users/idle` and `/users/idle?members=true` ### Link with Issue [#635](https://github.com/Real-Dev-Squad/website-backend/issues/635)
1.0
API Contract for GET Idle users/members - ### AC - Create an API contract for `/users/idle` and `/users/idle?members=true` ### Link with Issue [#635](https://github.com/Real-Dev-Squad/website-backend/issues/635)
non_process
api contract for get idle users members ac create an api contract for users idle and users idle members true link with issue
0
163
2,583,797,330
IssuesEvent
2015-02-16 10:17:32
luc-github/Repetier-Firmware-0.92
https://api.github.com/repos/luc-github/Repetier-Firmware-0.92
closed
Build fails on Arduino Nightly (1.6.0) Jan 14
enhancement Waiting to be processed
It looks like the current nightly builds of Arduino have incompatible changes with variants.cpp. I was able to build by copying the "hardware" directory from 1.5.8.
1.0
Build fails on Arduino Nightly (1.6.0) Jan 14 - It looks like the current nightly builds of Arduino have incompatible changes with variants.cpp. I was able to build by copying the "hardware" directory from 1.5.8.
process
build fails on arduino nightly jan it looks like the current nightly builds of arduino have incompatible changes with variants cpp i was able to build by copying the hardware directory from
1
4,394
7,285,884,163
IssuesEvent
2018-02-23 07:12:33
muflihun/residue
https://api.github.com/repos/muflihun/residue
closed
Remove plain log request support
area: log-processing type: improvement
This is a security concern + extra instructions for potentially unused feature We will remove it in 1.5.0
1.0
Remove plain log request support - This is a security concern + extra instructions for potentially unused feature We will remove it in 1.5.0
process
remove plain log request support this is a security concern extra instructions for potentially unused feature we will remove it in
1
115,529
14,799,011,046
IssuesEvent
2021-01-13 01:13:01
vmware-tanzu/antrea
https://api.github.com/repos/vmware-tanzu/antrea
opened
Replace hack/netpol/ with new upstream NetworkPolicy test suite
area/test/community kind/design priority/important-longterm
The netpol test suite that we protoyped in Antrea (https://github.com/vmware-tanzu/antrea/tree/master/hack/netpol) was ported upstream by @jayunit100 and others: https://github.com/kubernetes/kubernetes/tree/master/test/e2e/network/netpol As a result it now makes sense to remove the `hack/netpol/` directory altogether and instead start running the upstream version of the test suite as part of Antrea CI. I think we can probably move from a Kind CI job to a Jenkins job on VMC (VMware on AWS), to avoid some known issues with Kind / the Open vSwitch netdev datapath (see https://github.com/vmware-tanzu/antrea/issues/897 for an example). My preference would be to simply run it as part of the existing `jenkins-networkpolicy` job: * because it makes sense :) * to avoid introducing yet another job * because the test suite is supposed to run pretty fast and should not add too much time compared to the current job I am hoping we can simply update https://github.com/vmware-tanzu/antrea/blob/master/ci/run-k8s-e2e-tests.sh to avoid increasing the number of CI scripts we have to maintain. The only issue I see is that at this time (01/12), there is no named tag of the `k8s.gcr.io/conformance` image which includes the upstream netpol tests. The latest tag seems to be `v1.21.0-alpha.0`, and it doesn't include them. As a result, we can either: 1. wait for a version of the `k8s.gcr.io/conformance` image with support for the netpol tests 2. build the `k8s.gcr.io/conformance` image from the K8s source (https://github.com/kubernetes/kubernetes/blob/master/cluster/images/conformance/Makefile) and use it in `run-k8s-e2e-tests.sh` I don't really have a preference for either. If someone wants to start working on this issue in the near future, then they can go with the second solution.
1.0
Replace hack/netpol/ with new upstream NetworkPolicy test suite - The netpol test suite that we protoyped in Antrea (https://github.com/vmware-tanzu/antrea/tree/master/hack/netpol) was ported upstream by @jayunit100 and others: https://github.com/kubernetes/kubernetes/tree/master/test/e2e/network/netpol As a result it now makes sense to remove the `hack/netpol/` directory altogether and instead start running the upstream version of the test suite as part of Antrea CI. I think we can probably move from a Kind CI job to a Jenkins job on VMC (VMware on AWS), to avoid some known issues with Kind / the Open vSwitch netdev datapath (see https://github.com/vmware-tanzu/antrea/issues/897 for an example). My preference would be to simply run it as part of the existing `jenkins-networkpolicy` job: * because it makes sense :) * to avoid introducing yet another job * because the test suite is supposed to run pretty fast and should not add too much time compared to the current job I am hoping we can simply update https://github.com/vmware-tanzu/antrea/blob/master/ci/run-k8s-e2e-tests.sh to avoid increasing the number of CI scripts we have to maintain. The only issue I see is that at this time (01/12), there is no named tag of the `k8s.gcr.io/conformance` image which includes the upstream netpol tests. The latest tag seems to be `v1.21.0-alpha.0`, and it doesn't include them. As a result, we can either: 1. wait for a version of the `k8s.gcr.io/conformance` image with support for the netpol tests 2. build the `k8s.gcr.io/conformance` image from the K8s source (https://github.com/kubernetes/kubernetes/blob/master/cluster/images/conformance/Makefile) and use it in `run-k8s-e2e-tests.sh` I don't really have a preference for either. If someone wants to start working on this issue in the near future, then they can go with the second solution.
non_process
replace hack netpol with new upstream networkpolicy test suite the netpol test suite that we protoyped in antrea was ported upstream by and others as a result it now makes sense to remove the hack netpol directory altogether and instead start running the upstream version of the test suite as part of antrea ci i think we can probably move from a kind ci job to a jenkins job on vmc vmware on aws to avoid some known issues with kind the open vswitch netdev datapath see for an example my preference would be to simply run it as part of the existing jenkins networkpolicy job because it makes sense to avoid introducing yet another job because the test suite is supposed to run pretty fast and should not add too much time compared to the current job i am hoping we can simply update to avoid increasing the number of ci scripts we have to maintain the only issue i see is that at this time there is no named tag of the gcr io conformance image which includes the upstream netpol tests the latest tag seems to be alpha and it doesn t include them as a result we can either wait for a version of the gcr io conformance image with support for the netpol tests build the gcr io conformance image from the source and use it in run tests sh i don t really have a preference for either if someone wants to start working on this issue in the near future then they can go with the second solution
0
615,216
19,250,014,456
IssuesEvent
2021-12-09 03:17:19
matrixorigin/matrixone
https://api.github.com/repos/matrixorigin/matrixone
opened
add AOE RFC documents
component/aoe priority/high kind/feature severity/critical
1. Overall architecture 2. WAL 3. Buffer manager 4. Metadata 5. Data and index management 6. MVCC 7. Logstore
1.0
add AOE RFC documents - 1. Overall architecture 2. WAL 3. Buffer manager 4. Metadata 5. Data and index management 6. MVCC 7. Logstore
non_process
add aoe rfc documents overall architecture wal buffer manager metadata data and index management mvcc logstore
0
412,455
27,859,147,553
IssuesEvent
2023-03-21 03:34:54
CodeforHawaii/HIERR
https://api.github.com/repos/CodeforHawaii/HIERR
closed
Define user stories
documentation High Priority
This is for defining what user stories there are. Once this list is defined, we will send to Scott for prioritization.
1.0
Define user stories - This is for defining what user stories there are. Once this list is defined, we will send to Scott for prioritization.
non_process
define user stories this is for defining what user stories there are once this list is defined we will send to scott for prioritization
0
372,020
11,007,923,144
IssuesEvent
2019-12-04 09:31:36
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
t.co - see bug description
browser-focus-geckoview engine-gecko priority-critical
<!-- @browser: Firefox Mobile 71.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:71.0) Gecko/71.0 Firefox/71.0 --> <!-- @reported_with: --> <!-- @extra_labels: browser-focus-geckoview --> **URL**: https://t.co/YDbDt5izbJ?amp=1 **Browser / Version**: Firefox Mobile 71.0 **Operating System**: Android **Tested Another Browser**: Yes **Problem type**: Something else **Description**: Firefox focus does not trust user level https certificates, making Adguard https filtering broken **Steps to Reproduce**: Firefox focus does not trust user level https certificates, making Adguard https filtering broken <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
t.co - see bug description - <!-- @browser: Firefox Mobile 71.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:71.0) Gecko/71.0 Firefox/71.0 --> <!-- @reported_with: --> <!-- @extra_labels: browser-focus-geckoview --> **URL**: https://t.co/YDbDt5izbJ?amp=1 **Browser / Version**: Firefox Mobile 71.0 **Operating System**: Android **Tested Another Browser**: Yes **Problem type**: Something else **Description**: Firefox focus does not trust user level https certificates, making Adguard https filtering broken **Steps to Reproduce**: Firefox focus does not trust user level https certificates, making Adguard https filtering broken <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
t co see bug description url browser version firefox mobile operating system android tested another browser yes problem type something else description firefox focus does not trust user level https certificates making adguard https filtering broken steps to reproduce firefox focus does not trust user level https certificates making adguard https filtering broken browser configuration none from with ❤️
0
10,448
13,227,205,896
IssuesEvent
2020-08-18 02:25:35
googlemaps/v3-utility-library
https://api.github.com/repos/googlemaps/v3-utility-library
closed
markerclustererplus unit tests
help wanted priority: p3 stale type: process
recently had issues such as #549 that could have been caught through unittests
1.0
markerclustererplus unit tests - recently had issues such as #549 that could have been caught through unittests
process
markerclustererplus unit tests recently had issues such as that could have been caught through unittests
1
191
2,519,739,944
IssuesEvent
2015-01-18 09:00:56
mbunkus/mtx-trac-import-test
https://api.github.com/repos/mbunkus/mtx-trac-import-test
opened
mkvmerge will freeze if used with many tracks
C: mkvmerge P: normal R: fixed T: defect
**Reported by moritz on 9 Aug 2003 20:40 UTC** Liisachan reports: I tested 3 samples made by MatroskaMuxer (1) Xvid + 1 Vorbis + 5 ssa = 7 tracks (2) xvid + 1 vorbis + 16 ssa = 18 tracks (3) xvid + 1 vorbis + 17 ssa = 19 tracks Command line is like: mkvmerge -o "out.mkv" --language 1:rus --language 2:rus --language 3:eng --language 4:jpn --language 5:fre --language 6:dut --language 7:rus "in.mkv" > debug.txt mkvmerge 0.5.0 works fine for the all, while 0.6.0 works only for (1) 0.6.0 will freeze for (2) and (3) 0.6.0 can handle Chapter / Attachment files for (1) but 0.6.0 cannot handle (2) and (3) even without chapters etc... 0.6.0 doesnt't like too many tracks, i wonder? Migrated-From: https://trac.bunkus.org/ticket/8
1.0
mkvmerge will freeze if used with many tracks - **Reported by moritz on 9 Aug 2003 20:40 UTC** Liisachan reports: I tested 3 samples made by MatroskaMuxer (1) Xvid + 1 Vorbis + 5 ssa = 7 tracks (2) xvid + 1 vorbis + 16 ssa = 18 tracks (3) xvid + 1 vorbis + 17 ssa = 19 tracks Command line is like: mkvmerge -o "out.mkv" --language 1:rus --language 2:rus --language 3:eng --language 4:jpn --language 5:fre --language 6:dut --language 7:rus "in.mkv" > debug.txt mkvmerge 0.5.0 works fine for the all, while 0.6.0 works only for (1) 0.6.0 will freeze for (2) and (3) 0.6.0 can handle Chapter / Attachment files for (1) but 0.6.0 cannot handle (2) and (3) even without chapters etc... 0.6.0 doesnt't like too many tracks, i wonder? Migrated-From: https://trac.bunkus.org/ticket/8
non_process
mkvmerge will freeze if used with many tracks reported by moritz on aug utc liisachan reports i tested samples made by matroskamuxer xvid vorbis ssa tracks xvid vorbis ssa tracks xvid vorbis ssa tracks command line is like mkvmerge o out mkv language rus language rus language eng language jpn language fre language dut language rus in mkv debug txt mkvmerge works fine for the all while works only for will freeze for and can handle chapter attachment files for but cannot handle and even without chapters etc doesnt t like too many tracks i wonder migrated from
0
12,839
15,223,150,502
IssuesEvent
2021-02-18 01:58:51
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Allow tool output and export to save a new layer/table in existing databases also for non QGIS native tools.
Feature Request Feedback Processing stale
First, it'd be nice to be able to save the output of a tool directly to a PostGIS layer/table. It seems that the only way to do this right now is to dump it out to a different format (shapefile/geojson) and then import it to PostGIS. Second, it would be nice to save tool output as a new layer in an existing database. Many tools can save layers to a **new** geopackage or spatialite database, but it doesn't seem possible to save the output as a new layer/table in an **existing** database. I like keeping all of my layers for a project in a single database - that makes it easier to move around. And I also use SQL in a lot of my analyses -it's fairly clunky to do joins between tables in separate databases. Thanks to all of the QGIS developers out there! I'm more and more impressed with every release!
1.0
Allow tool output and export to save a new layer/table in existing databases also for non QGIS native tools. - First, it'd be nice to be able to save the output of a tool directly to a PostGIS layer/table. It seems that the only way to do this right now is to dump it out to a different format (shapefile/geojson) and then import it to PostGIS. Second, it would be nice to save tool output as a new layer in an existing database. Many tools can save layers to a **new** geopackage or spatialite database, but it doesn't seem possible to save the output as a new layer/table in an **existing** database. I like keeping all of my layers for a project in a single database - that makes it easier to move around. And I also use SQL in a lot of my analyses -it's fairly clunky to do joins between tables in separate databases. Thanks to all of the QGIS developers out there! I'm more and more impressed with every release!
process
allow tool output and export to save a new layer table in existing databases also for non qgis native tools first it d be nice to be able to save the output of a tool directly to a postgis layer table it seems that the only way to do this right now is to dump it out to a different format shapefile geojson and then import it to postgis second it would be nice to save tool output as a new layer in an existing database many tools can save layers to a new geopackage or spatialite database but it doesn t seem possible to save the output as a new layer table in an existing database i like keeping all of my layers for a project in a single database that makes it easier to move around and i also use sql in a lot of my analyses it s fairly clunky to do joins between tables in separate databases thanks to all of the qgis developers out there i m more and more impressed with every release
1
18,119
24,150,759,854
IssuesEvent
2022-09-22 00:17:44
allinurl/goaccess
https://api.github.com/repos/allinurl/goaccess
closed
Parse Log with hostname instead of ipv4/ipv6
question log-processing command-line options
I have a log-file where the ip is already resolved and instead of ip has a hostname as a field. It seems this field is not understood by goaccess. As I understand the hostname field is manadatory, but is it really a requirement to hold an IP? `Token 'some.perfect-valid-hostname.example.com' doesn't match specifier '%h'` Does someone have an advice?
1.0
Parse Log with hostname instead of ipv4/ipv6 - I have a log-file where the ip is already resolved and instead of ip has a hostname as a field. It seems this field is not understood by goaccess. As I understand the hostname field is manadatory, but is it really a requirement to hold an IP? `Token 'some.perfect-valid-hostname.example.com' doesn't match specifier '%h'` Does someone have an advice?
process
parse log with hostname instead of i have a log file where the ip is already resolved and instead of ip has a hostname as a field it seems this field is not understood by goaccess as i understand the hostname field is manadatory but is it really a requirement to hold an ip token some perfect valid hostname example com doesn t match specifier h does someone have an advice
1
42,870
7,006,737,191
IssuesEvent
2017-12-19 09:39:54
rkt/rkt
https://api.github.com/repos/rkt/rkt
closed
Describe how to run unit/functional tests in the hacking guide.
area/developer tooling kind/cleanup kind/documentation
To make new developers' life easier, we need to document/link the guide on how to run unit/functional tests locally.
1.0
Describe how to run unit/functional tests in the hacking guide. - To make new developers' life easier, we need to document/link the guide on how to run unit/functional tests locally.
non_process
describe how to run unit functional tests in the hacking guide to make new developers life easier we need to document link the guide on how to run unit functional tests locally
0
5,304
8,124,297,086
IssuesEvent
2018-08-16 17:05:12
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
Intermittent: Empty Process stdout
area-System.Diagnostics.Process needs more info question
## Story I don't know why, but sometimes when running same console program with same input data, System.Diagnostics fucntionality returns empty stdout. It's not a problem with running process. Also I checked my code many times, no errors found. ## Platforms - Windows (tested) - Linux (tested) ## I think this non-repro code would be useful for creation of working repro of this problem I tried to do a minimal repro, but all tests passed and none failed... so maybe something wrong with my code. Or, if someone have same problem, maybe something wrong in .NET Core. ````c# using System; using System.IO; using System.Text; using System.Threading; using System.Diagnostics; class EmptyStdOutRepro { public static void Main() { new EmptyStdOutRepro().Run(); } public void Run() { while (true) Exec(); } private void Exec() { int lineCount = 0; string output = ""; Process process = new Process(); process.StartInfo.FileName = "ipconfig.exe"; process.StartInfo.UseShellExecute = false; process.StartInfo.RedirectStandardOutput = true; process.StartInfo.CreateNoWindow = true; process.StartInfo.ErrorDialog = false; process.OutputDataReceived += new DataReceivedEventHandler((sender, e) => { if (!String.IsNullOrEmpty(e.Data)) { lineCount++; output += ("\n[" + lineCount + "]: " + e.Data); } }); process.Start(); process.BeginOutputReadLine(); process.WaitForExit(); if (output != "") Console.WriteLine("PASSED!" + output.Length); else Console.WriteLine("FAILED!" + output.Length); process.WaitForExit(); process.Close(); } } ```` ## Working repro, needs minimization Class ProgramExecutor in SimplePM_Server repo: https://github.com/SirkadirovTeam/SimplePM_Server/tree/master/SimplePM_Server/ProgramTesting/SRunner/ProgramExecutor
1.0
Intermittent: Empty Process stdout - ## Story I don't know why, but sometimes when running same console program with same input data, System.Diagnostics fucntionality returns empty stdout. It's not a problem with running process. Also I checked my code many times, no errors found. ## Platforms - Windows (tested) - Linux (tested) ## I think this non-repro code would be useful for creation of working repro of this problem I tried to do a minimal repro, but all tests passed and none failed... so maybe something wrong with my code. Or, if someone have same problem, maybe something wrong in .NET Core. ````c# using System; using System.IO; using System.Text; using System.Threading; using System.Diagnostics; class EmptyStdOutRepro { public static void Main() { new EmptyStdOutRepro().Run(); } public void Run() { while (true) Exec(); } private void Exec() { int lineCount = 0; string output = ""; Process process = new Process(); process.StartInfo.FileName = "ipconfig.exe"; process.StartInfo.UseShellExecute = false; process.StartInfo.RedirectStandardOutput = true; process.StartInfo.CreateNoWindow = true; process.StartInfo.ErrorDialog = false; process.OutputDataReceived += new DataReceivedEventHandler((sender, e) => { if (!String.IsNullOrEmpty(e.Data)) { lineCount++; output += ("\n[" + lineCount + "]: " + e.Data); } }); process.Start(); process.BeginOutputReadLine(); process.WaitForExit(); if (output != "") Console.WriteLine("PASSED!" + output.Length); else Console.WriteLine("FAILED!" + output.Length); process.WaitForExit(); process.Close(); } } ```` ## Working repro, needs minimization Class ProgramExecutor in SimplePM_Server repo: https://github.com/SirkadirovTeam/SimplePM_Server/tree/master/SimplePM_Server/ProgramTesting/SRunner/ProgramExecutor
process
intermittent empty process stdout story i don t know why but sometimes when running same console program with same input data system diagnostics fucntionality returns empty stdout it s not a problem with running process also i checked my code many times no errors found platforms windows tested linux tested i think this non repro code would be useful for creation of working repro of this problem i tried to do a minimal repro but all tests passed and none failed so maybe something wrong with my code or if someone have same problem maybe something wrong in net core c using system using system io using system text using system threading using system diagnostics class emptystdoutrepro public static void main new emptystdoutrepro run public void run while true exec private void exec int linecount string output process process new process process startinfo filename ipconfig exe process startinfo useshellexecute false process startinfo redirectstandardoutput true process startinfo createnowindow true process startinfo errordialog false process outputdatareceived new datareceivedeventhandler sender e if string isnullorempty e data linecount output n e data process start process beginoutputreadline process waitforexit if output console writeline passed output length else console writeline failed output length process waitforexit process close working repro needs minimization class programexecutor in simplepm server repo
1
331,247
28,760,770,983
IssuesEvent
2023-05-01 00:03:24
WordPress/gutenberg
https://api.github.com/repos/WordPress/gutenberg
closed
[Flaky Test] User Mention: should allow newlines after multiple mention completions
[Status] Stale [Type] Flaky Test
<!-- __META_DATA__:{} --> **Flaky test detected. This is an auto-generated issue by GitHub Actions. Please do NOT edit this manually.** ## Test title User Mention: should allow newlines after multiple mention completions ## Test path `/test/e2e/specs/editor/various/autocomplete-and-mentions.spec.js` ## Errors <!-- __TEST_RESULTS_LIST__ --> <!-- __TEST_RESULT__ --><time datetime="2023-02-27T19:57:03.719Z"><code>[2023-02-27T19:57:03.719Z]</code></time> Test passed after 5 failed attempts on <a href="https://github.com/WordPress/gutenberg/actions/runs/4285345332"><code>try/set-attrs-idle</code></a>.<!-- /__TEST_RESULT__ --> <br/> <!-- __TEST_RESULT__ --><details> <summary> <time datetime="2023-03-30T07:42:51.930Z"><code>[2023-03-30T07:42:51.930Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/4561612597"><code>fix/padding-input-in-global-styles</code></a>. </summary> ``` Error: keyboard.type: Target page, context or browser has been closed at /home/runner/work/gutenberg/gutenberg/test/e2e/specs/editor/various/autocomplete-and-mentions.spec.js:403:25 ``` </details><!-- /__TEST_RESULT__ --> <!-- /__TEST_RESULTS_LIST__ -->
1.0
[Flaky Test] User Mention: should allow newlines after multiple mention completions - <!-- __META_DATA__:{} --> **Flaky test detected. This is an auto-generated issue by GitHub Actions. Please do NOT edit this manually.** ## Test title User Mention: should allow newlines after multiple mention completions ## Test path `/test/e2e/specs/editor/various/autocomplete-and-mentions.spec.js` ## Errors <!-- __TEST_RESULTS_LIST__ --> <!-- __TEST_RESULT__ --><time datetime="2023-02-27T19:57:03.719Z"><code>[2023-02-27T19:57:03.719Z]</code></time> Test passed after 5 failed attempts on <a href="https://github.com/WordPress/gutenberg/actions/runs/4285345332"><code>try/set-attrs-idle</code></a>.<!-- /__TEST_RESULT__ --> <br/> <!-- __TEST_RESULT__ --><details> <summary> <time datetime="2023-03-30T07:42:51.930Z"><code>[2023-03-30T07:42:51.930Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/4561612597"><code>fix/padding-input-in-global-styles</code></a>. </summary> ``` Error: keyboard.type: Target page, context or browser has been closed at /home/runner/work/gutenberg/gutenberg/test/e2e/specs/editor/various/autocomplete-and-mentions.spec.js:403:25 ``` </details><!-- /__TEST_RESULT__ --> <!-- /__TEST_RESULTS_LIST__ -->
non_process
user mention should allow newlines after multiple mention completions flaky test detected this is an auto generated issue by github actions please do not edit this manually test title user mention should allow newlines after multiple mention completions test path test specs editor various autocomplete and mentions spec js errors test passed after failed attempts on test passed after failed attempt on a href error keyboard type target page context or browser has been closed at home runner work gutenberg gutenberg test specs editor various autocomplete and mentions spec js
0
22,325
30,913,247,439
IssuesEvent
2023-08-05 01:26:45
nodejs/node
https://api.github.com/repos/nodejs/node
closed
child_process,tty: pseudo-tty support
child_process feature request libuv blocked stale
Blocked on someone implementing https://github.com/libuv/libuv/issues/2640, then exposing it in node. Refs #31409 and others.
1.0
child_process,tty: pseudo-tty support - Blocked on someone implementing https://github.com/libuv/libuv/issues/2640, then exposing it in node. Refs #31409 and others.
process
child process tty pseudo tty support blocked on someone implementing then exposing it in node refs and others
1
111,587
9,535,889,137
IssuesEvent
2019-04-30 08:17:56
chameleon-system/chameleon-system
https://api.github.com/repos/chameleon-system/chameleon-system
closed
Image Sharpening fails on small images with image magick 6.9.x
Priority: High Status: Test Type: Bug
**Describe the bug** When sharpening images is enabled for all image sizes, then thumbnail creation will fail for smaller images (example image size tested was 59x26) **Affected version(s)** Affects all chameleon versions **To Reproduce** Steps to reproduce the behavior: 1. In the cms backend go to cms settings and open the tab images 2. Add an entry without a size limitation and sharpening enabled 3. remove all thumbnails and clear cache 4. attempt to view the image via frontend **Expected behavior** A sharpend version of the image is generated. Instead, an error will be show in dev mode. in prod mode, no error message is shown but the image is not generated. **Technical details** source of the problem appears to be the parameter `-unsharp 1.5×1.0+1.5+0.02` being added to the imagemagick command in `\imageMagick::ResizeImage`. Executing the same command directly on the command line will produce the following error: ``` convert: invalid argument for option `-unsharp': 1.5×1.0+1.5+0.02 @ error/convert.c/ConvertImageCommand/3128. ``` The error occurs with imagemagick 6.9.7-4 Q16 x86_64 2017011 but not with version 6.8.9-9 Q16 x86_64 2018-11-11 Replacing `-unsharp 1.5×1.0+1.5+0.02` with `-adaptive-sharpen 0x1` appears to produce similar sharpness results and works for small images. This may be an alternative.
1.0
Image Sharpening fails on small images with image magick 6.9.x - **Describe the bug** When sharpening images is enabled for all image sizes, then thumbnail creation will fail for smaller images (example image size tested was 59x26) **Affected version(s)** Affects all chameleon versions **To Reproduce** Steps to reproduce the behavior: 1. In the cms backend go to cms settings and open the tab images 2. Add an entry without a size limitation and sharpening enabled 3. remove all thumbnails and clear cache 4. attempt to view the image via frontend **Expected behavior** A sharpend version of the image is generated. Instead, an error will be show in dev mode. in prod mode, no error message is shown but the image is not generated. **Technical details** source of the problem appears to be the parameter `-unsharp 1.5×1.0+1.5+0.02` being added to the imagemagick command in `\imageMagick::ResizeImage`. Executing the same command directly on the command line will produce the following error: ``` convert: invalid argument for option `-unsharp': 1.5×1.0+1.5+0.02 @ error/convert.c/ConvertImageCommand/3128. ``` The error occurs with imagemagick 6.9.7-4 Q16 x86_64 2017011 but not with version 6.8.9-9 Q16 x86_64 2018-11-11 Replacing `-unsharp 1.5×1.0+1.5+0.02` with `-adaptive-sharpen 0x1` appears to produce similar sharpness results and works for small images. This may be an alternative.
non_process
image sharpening fails on small images with image magick x describe the bug when sharpening images is enabled for all image sizes then thumbnail creation will fail for smaller images example image size tested was affected version s affects all chameleon versions to reproduce steps to reproduce the behavior in the cms backend go to cms settings and open the tab images add an entry without a size limitation and sharpening enabled remove all thumbnails and clear cache attempt to view the image via frontend expected behavior a sharpend version of the image is generated instead an error will be show in dev mode in prod mode no error message is shown but the image is not generated technical details source of the problem appears to be the parameter unsharp × being added to the imagemagick command in imagemagick resizeimage executing the same command directly on the command line will produce the following error convert invalid argument for option unsharp × error convert c convertimagecommand the error occurs with imagemagick but not with version replacing unsharp × with adaptive sharpen appears to produce similar sharpness results and works for small images this may be an alternative
0
17,575
23,387,109,980
IssuesEvent
2022-08-11 14:35:57
vectordotdev/vector
https://api.github.com/repos/vectordotdev/vector
closed
New `recollect_fields` transform
needs: approval domain: transforms type: feature domain: processing
It would be nice to have a transform that, if a given field is not present in the event, can insert this field with last observed value of this field from previous events. Example configuration: ```toml [transforms.recollect_fields] inputs = ["in"] fields = ["field1", "field2"] ``` Such a transform could have been implemented by a user with a stateful JavaScript transform (#721) as ```javascript const fields = ["field1", "field2"]; const storedValues = {}; const handler = event => { for (const field of fields) { if (event.hasOwnProperty(field)) { storedValues[field] = event[field]; } else if (storedValues.hasOwnProperty(field)) { event[field] = storedValues[field]; } } return event; } ``` As a bonus, this transform would allow to join multiple streams of events. For example, for this config ```toml [inputs.cpu_temp] # produces events with field "cpu_temp" [inputs.cpu_freq] # produces events with field "cpu_freq" [transforms.cpu_state] type = "recollect_fields" inputs = ["cpu_temp", "cpu_freq"] fields = ["cpu_temp", "cpu_freq"] [outputs.console] type = "console" inputs = ["recollect_fields"] ``` events containing both temperature and frequency would be printed. In addition to the most basic form described above, we can add an optional configuration parameter `forget_time` which would prevent the field value from being recollected if it was produced earlier than the the current time minus timeout.
1.0
New `recollect_fields` transform - It would be nice to have a transform that, if a given field is not present in the event, can insert this field with last observed value of this field from previous events. Example configuration: ```toml [transforms.recollect_fields] inputs = ["in"] fields = ["field1", "field2"] ``` Such a transform could have been implemented by a user with a stateful JavaScript transform (#721) as ```javascript const fields = ["field1", "field2"]; const storedValues = {}; const handler = event => { for (const field of fields) { if (event.hasOwnProperty(field)) { storedValues[field] = event[field]; } else if (storedValues.hasOwnProperty(field)) { event[field] = storedValues[field]; } } return event; } ``` As a bonus, this transform would allow to join multiple streams of events. For example, for this config ```toml [inputs.cpu_temp] # produces events with field "cpu_temp" [inputs.cpu_freq] # produces events with field "cpu_freq" [transforms.cpu_state] type = "recollect_fields" inputs = ["cpu_temp", "cpu_freq"] fields = ["cpu_temp", "cpu_freq"] [outputs.console] type = "console" inputs = ["recollect_fields"] ``` events containing both temperature and frequency would be printed. In addition to the most basic form described above, we can add an optional configuration parameter `forget_time` which would prevent the field value from being recollected if it was produced earlier than the the current time minus timeout.
process
new recollect fields transform it would be nice to have a transform that if a given field is not present in the event can insert this field with last observed value of this field from previous events example configuration toml inputs fields such a transform could have been implemented by a user with a stateful javascript transform as javascript const fields const storedvalues const handler event for const field of fields if event hasownproperty field storedvalues event else if storedvalues hasownproperty field event storedvalues return event as a bonus this transform would allow to join multiple streams of events for example for this config toml produces events with field cpu temp produces events with field cpu freq type recollect fields inputs fields type console inputs events containing both temperature and frequency would be printed in addition to the most basic form described above we can add an optional configuration parameter forget time which would prevent the field value from being recollected if it was produced earlier than the the current time minus timeout
1
11,571
14,441,672,673
IssuesEvent
2020-12-07 17:05:34
frontendbr/forum
https://api.github.com/repos/frontendbr/forum
closed
Como medir produtividade?
Processos [Discussão]
Você trabalha home office ou no escritório? Freela ou não? Como medir sua produtividade com tantas variações de modos de trabalho por ai? Como vocês se avaliam?
1.0
Como medir produtividade? - Você trabalha home office ou no escritório? Freela ou não? Como medir sua produtividade com tantas variações de modos de trabalho por ai? Como vocês se avaliam?
process
como medir produtividade você trabalha home office ou no escritório freela ou não como medir sua produtividade com tantas variações de modos de trabalho por ai como vocês se avaliam
1
590
3,067,143,898
IssuesEvent
2015-08-18 08:35:47
maraujop/django-crispy-forms
https://api.github.com/repos/maraujop/django-crispy-forms
closed
Use py.test for testing
Testing/Process
Hi! I want to propose to rewrite test suite with py.test. Main goals are: * Simplicity. Now test suite is quite complicated and has a lot of code duplication * Granularity. py.test has beautiful feature to [mark tests](http://pytest.org/latest/parametrize.html#parametrized-test-functions) to run it in different circumstances. Running tests for different templates pack would be easy. Also py.test has [parametrization](http://pytest.org/latest/parametrize.html#parametrized-test-functions), so it should plays great with different templates packs too. * Split setup and tests. Test code would be more clear and will have less duplication. Also test [fixtures](http://pytest.org/latest/fixture.html#fixture) will be reusable. It will be better to show the code, so, don't you mind if I'll create some draft for that?
1.0
Use py.test for testing - Hi! I want to propose to rewrite test suite with py.test. Main goals are: * Simplicity. Now test suite is quite complicated and has a lot of code duplication * Granularity. py.test has beautiful feature to [mark tests](http://pytest.org/latest/parametrize.html#parametrized-test-functions) to run it in different circumstances. Running tests for different templates pack would be easy. Also py.test has [parametrization](http://pytest.org/latest/parametrize.html#parametrized-test-functions), so it should plays great with different templates packs too. * Split setup and tests. Test code would be more clear and will have less duplication. Also test [fixtures](http://pytest.org/latest/fixture.html#fixture) will be reusable. It will be better to show the code, so, don't you mind if I'll create some draft for that?
process
use py test for testing hi i want to propose to rewrite test suite with py test main goals are simplicity now test suite is quite complicated and has a lot of code duplication granularity py test has beautiful feature to to run it in different circumstances running tests for different templates pack would be easy also py test has so it should plays great with different templates packs too split setup and tests test code would be more clear and will have less duplication also test will be reusable it will be better to show the code so don t you mind if i ll create some draft for that
1
43,318
7,042,094,892
IssuesEvent
2017-12-30 06:30:25
zulip/zulip
https://api.github.com/repos/zulip/zulip
closed
user docs: Clarify some items
area: documentation (user) enhancement
I've noticed a few broken links or inconsistencies in the docs, leading to some possibly unclear parts of the documentation. - [ ] https://chat.zulip.org/help/getting-started-with-zulip. At the bottom of this page (under `Next Steps`), the **"Add an avatar"** and **"Configure your notifications"** links link directly to the settings, unlike many of the other links in the documentation, which link you to the page that shows how to do those things. - [ ] https://chat.zulip.org/help/change-your-password. For number 2 under "If you forgot your password", the terminology on the Zulip homepage has been changed to "Forgot your password?" from "Forgot password?" - [ ] https://chat.zulip.org/help/signing-in. There are more than 3 ways to sign in that aren't documented here. - [ ] https://chat.zulip.org/help/analytics. This page should be a little bit more clear on how to get to the statistics; it's fairly easy to miss the one (broken) link at the top of the page. - [ ] https://chat.zulip.org/help/zulip-on-windows. This and the page for Android should be deleted as they seem to be from older versions of those apps. The link for the Windows download is broken on this page, but the Android one links to the correct Google Play Store listing. - [ ] https://chat.zulip.org/help/about-streams-and-topics. On this page, for the links to customize stream settings, there are many items missing, so something should be added to lead users to the other customization options. - [ ] https://chat.zulip.org/help/change-your-organization-settings. This one is extremely minor, but an image for a chevron is present next to all the other instances of the word "chevron" in the docs (same with "gear"), but it's missing here.
1.0
user docs: Clarify some items - I've noticed a few broken links or inconsistencies in the docs, leading to some possibly unclear parts of the documentation. - [ ] https://chat.zulip.org/help/getting-started-with-zulip. At the bottom of this page (under `Next Steps`), the **"Add an avatar"** and **"Configure your notifications"** links link directly to the settings, unlike many of the other links in the documentation, which link you to the page that shows how to do those things. - [ ] https://chat.zulip.org/help/change-your-password. For number 2 under "If you forgot your password", the terminology on the Zulip homepage has been changed to "Forgot your password?" from "Forgot password?" - [ ] https://chat.zulip.org/help/signing-in. There are more than 3 ways to sign in that aren't documented here. - [ ] https://chat.zulip.org/help/analytics. This page should be a little bit more clear on how to get to the statistics; it's fairly easy to miss the one (broken) link at the top of the page. - [ ] https://chat.zulip.org/help/zulip-on-windows. This and the page for Android should be deleted as they seem to be from older versions of those apps. The link for the Windows download is broken on this page, but the Android one links to the correct Google Play Store listing. - [ ] https://chat.zulip.org/help/about-streams-and-topics. On this page, for the links to customize stream settings, there are many items missing, so something should be added to lead users to the other customization options. - [ ] https://chat.zulip.org/help/change-your-organization-settings. This one is extremely minor, but an image for a chevron is present next to all the other instances of the word "chevron" in the docs (same with "gear"), but it's missing here.
non_process
user docs clarify some items i ve noticed a few broken links or inconsistencies in the docs leading to some possibly unclear parts of the documentation at the bottom of this page under next steps the add an avatar and configure your notifications links link directly to the settings unlike many of the other links in the documentation which link you to the page that shows how to do those things for number under if you forgot your password the terminology on the zulip homepage has been changed to forgot your password from forgot password there are more than ways to sign in that aren t documented here this page should be a little bit more clear on how to get to the statistics it s fairly easy to miss the one broken link at the top of the page this and the page for android should be deleted as they seem to be from older versions of those apps the link for the windows download is broken on this page but the android one links to the correct google play store listing on this page for the links to customize stream settings there are many items missing so something should be added to lead users to the other customization options this one is extremely minor but an image for a chevron is present next to all the other instances of the word chevron in the docs same with gear but it s missing here
0
31,036
8,641,807,119
IssuesEvent
2018-11-24 21:54:19
Trustroots/trustroots
https://api.github.com/repos/Trustroots/trustroots
closed
easy installation for devs
build
It seems currently the only way to get this set up is if you have a _lot_ of knowledge of node, docker, and mongo. It should be possible to set this up in less than 15 minutes with minimal knowledge of any of these, at least on Debian/Ubuntu and OSX. How can we do this?
1.0
easy installation for devs - It seems currently the only way to get this set up is if you have a _lot_ of knowledge of node, docker, and mongo. It should be possible to set this up in less than 15 minutes with minimal knowledge of any of these, at least on Debian/Ubuntu and OSX. How can we do this?
non_process
easy installation for devs it seems currently the only way to get this set up is if you have a lot of knowledge of node docker and mongo it should be possible to set this up in less than minutes with minimal knowledge of any of these at least on debian ubuntu and osx how can we do this
0
299,815
22,625,193,509
IssuesEvent
2022-06-30 10:00:25
C3ETH/c3eth-operations
https://api.github.com/repos/C3ETH/c3eth-operations
closed
Catalyst [Challenge Team]: Attending Challenge Team Roundtable 5 meeting
Documentation:Catalyst Fund Catalyst:Fund 8 Treasury:Tasks
**Tasks:** - [x] Attend the Roundtable 5 on Friday, 06 May 2022 - [x] Check this [Fund8 Challenge Teams & Onboarding Guidance](https://docs.google.com/document/d/1lmlrqrTthz1CjCj2zNtY7hlAfGsKui_cXIUNygLPsm0/edit) - [x] Find two available 1.5 hours slot during 18-24 May 2022 for Proof of Life meetings - [x] Fill this [Google Form](https://docs.google.com/forms/d/e/1FAIpQLSfKrCwNypGi9vT5rCy2eaHKpe-PbKokFp7nftY_OuzhQmxSrg/viewform) for the slot that you want - [x] Confirmed at least two members of the CT will take part in the Proof of Life meetings - [x] [Miro board](https://miro.com/app/board/uXjVO4r7A_c=/)
1.0
Catalyst [Challenge Team]: Attending Challenge Team Roundtable 5 meeting - **Tasks:** - [x] Attend the Roundtable 5 on Friday, 06 May 2022 - [x] Check this [Fund8 Challenge Teams & Onboarding Guidance](https://docs.google.com/document/d/1lmlrqrTthz1CjCj2zNtY7hlAfGsKui_cXIUNygLPsm0/edit) - [x] Find two available 1.5 hours slot during 18-24 May 2022 for Proof of Life meetings - [x] Fill this [Google Form](https://docs.google.com/forms/d/e/1FAIpQLSfKrCwNypGi9vT5rCy2eaHKpe-PbKokFp7nftY_OuzhQmxSrg/viewform) for the slot that you want - [x] Confirmed at least two members of the CT will take part in the Proof of Life meetings - [x] [Miro board](https://miro.com/app/board/uXjVO4r7A_c=/)
non_process
catalyst attending challenge team roundtable meeting tasks attend the roundtable on friday may check this find two available hours slot during may for proof of life meetings fill this for the slot that you want confirmed at least two members of the ct will take part in the proof of life meetings
0
308,613
23,257,415,216
IssuesEvent
2022-08-04 10:29:53
neogulcity/jsonparser
https://api.github.com/repos/neogulcity/jsonparser
opened
데이터 타입 판단
documentation
```c++ int myInt[2]; int* myInt = new int[2]; ``` int[], int* 두 가지 데이터 타입에 따른 Json 배열 처리 방법이 필요함.
1.0
데이터 타입 판단 - ```c++ int myInt[2]; int* myInt = new int[2]; ``` int[], int* 두 가지 데이터 타입에 따른 Json 배열 처리 방법이 필요함.
non_process
데이터 타입 판단 c int myint int myint new int int int 두 가지 데이터 타입에 따른 json 배열 처리 방법이 필요함
0
4,550
7,375,395,204
IssuesEvent
2018-03-14 00:13:09
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Consistency in <acrName>
assigned-to-author container-service doc-enhancement in-process triaged
In the previous steps (1 and 2), you use <acrName> as a variable. In the below it's "myACRRegistry". Suggest to replace by <acrName> ACR_ID=$(az acr show --name myACRRegistry --resource-group myResourceGroup --query "id" --output tsv) --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: d4af71a4-648e-580f-7e86-94ce05196de5 * Version Independent ID: 134b8ba0-b085-076a-d2e1-16d0ddff5586 * Content: [Kubernetes on Azure tutorial - Deploy Cluster | Microsoft Docs](https://docs.microsoft.com/en-us/azure/aks/tutorial-kubernetes-deploy-cluster) * Content Source: [articles/aks/tutorial-kubernetes-deploy-cluster.md](https://github.com/Microsoft/azure-docs/blob/master/articles/aks/tutorial-kubernetes-deploy-cluster.md) * Service: **container-service** * GitHub Login: @neilpeterson * Microsoft Alias: **nepeters**
1.0
Consistency in <acrName> - In the previous steps (1 and 2), you use <acrName> as a variable. In the below it's "myACRRegistry". Suggest to replace by <acrName> ACR_ID=$(az acr show --name myACRRegistry --resource-group myResourceGroup --query "id" --output tsv) --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: d4af71a4-648e-580f-7e86-94ce05196de5 * Version Independent ID: 134b8ba0-b085-076a-d2e1-16d0ddff5586 * Content: [Kubernetes on Azure tutorial - Deploy Cluster | Microsoft Docs](https://docs.microsoft.com/en-us/azure/aks/tutorial-kubernetes-deploy-cluster) * Content Source: [articles/aks/tutorial-kubernetes-deploy-cluster.md](https://github.com/Microsoft/azure-docs/blob/master/articles/aks/tutorial-kubernetes-deploy-cluster.md) * Service: **container-service** * GitHub Login: @neilpeterson * Microsoft Alias: **nepeters**
process
consistency in in the previous steps and you use as a variable in the below it s myacrregistry suggest to replace by acr id az acr show name myacrregistry resource group myresourcegroup query id output tsv document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service container service github login neilpeterson microsoft alias nepeters
1
19,460
25,752,788,667
IssuesEvent
2022-12-08 14:23:56
influxdata/telegraf
https://api.github.com/repos/influxdata/telegraf
closed
processors.parser changes measurement name to 'parser'
bug plugin/parser plugin/processor
### Relevant telegraf.conf ```toml [[inputs.file]] files = ["/test/test.txt"] data_format = "value" data_type = "string" [[processors.parser]] namepass = ["file"] parse_fields = ["value"] drop_original = false merge = "override" data_format = "value" data_type = "float" [[outputs.file]] files = ["stdout"] ``` ### Logs from Telegraf ```text Telegraf 1.23.4: file value=7.2 1664873970000000000 Telegraf 1.24.2: parser value=7.2 1664892267000000000 ``` ### Steps to reproduce Use the [[processors.parser]] as shown in the example above or as reported in one of links given in the "Additional info" block. ### Expected behavior like Telegraf 1.23.4 does: keep the measurement name of the input. ### Actual behavior Measurement name is changed to 'parser'. The field is dropped from the original measurement, even with setting `drop_original = false`. ### Additional info The same issue get's talked about behind these 3 links: * https://community.influxdata.com/t/parser-processor-change-measurement-name-to-parser/26984 * https://community.influxdata.com/t/single-mqtt-consumer-for-int-float-string/27137 * https://influxcommunity.slack.com/archives/CH99HUH8V/p1666794045416509 I'm a new user of InfluxDB & Telegraf, and have no idea how to actually switch to an older Telegraf version. I've copied the sample and it's output from this issue: https://github.com/influxdata/telegraf/issues/11921
1.0
processors.parser changes measurement name to 'parser' - ### Relevant telegraf.conf ```toml [[inputs.file]] files = ["/test/test.txt"] data_format = "value" data_type = "string" [[processors.parser]] namepass = ["file"] parse_fields = ["value"] drop_original = false merge = "override" data_format = "value" data_type = "float" [[outputs.file]] files = ["stdout"] ``` ### Logs from Telegraf ```text Telegraf 1.23.4: file value=7.2 1664873970000000000 Telegraf 1.24.2: parser value=7.2 1664892267000000000 ``` ### Steps to reproduce Use the [[processors.parser]] as shown in the example above or as reported in one of links given in the "Additional info" block. ### Expected behavior like Telegraf 1.23.4 does: keep the measurement name of the input. ### Actual behavior Measurement name is changed to 'parser'. The field is dropped from the original measurement, even with setting `drop_original = false`. ### Additional info The same issue get's talked about behind these 3 links: * https://community.influxdata.com/t/parser-processor-change-measurement-name-to-parser/26984 * https://community.influxdata.com/t/single-mqtt-consumer-for-int-float-string/27137 * https://influxcommunity.slack.com/archives/CH99HUH8V/p1666794045416509 I'm a new user of InfluxDB & Telegraf, and have no idea how to actually switch to an older Telegraf version. I've copied the sample and it's output from this issue: https://github.com/influxdata/telegraf/issues/11921
process
processors parser changes measurement name to parser relevant telegraf conf toml files data format value data type string namepass parse fields drop original false merge override data format value data type float files logs from telegraf text telegraf file value telegraf parser value steps to reproduce use the as shown in the example above or as reported in one of links given in the additional info block expected behavior like telegraf does keep the measurement name of the input actual behavior measurement name is changed to parser the field is dropped from the original measurement even with setting drop original false additional info the same issue get s talked about behind these links i m a new user of influxdb telegraf and have no idea how to actually switch to an older telegraf version i ve copied the sample and it s output from this issue
1
260,934
8,221,308,300
IssuesEvent
2018-09-06 01:13:06
C3DSU/e-DefPR
https://api.github.com/repos/C3DSU/e-DefPR
closed
Padronizar nomes de atributos do backend
Category: Backend Priority: Medium Type: Bug
Padronizar nomes de atributos do backend, não utilizar mais CamelCase
1.0
Padronizar nomes de atributos do backend - Padronizar nomes de atributos do backend, não utilizar mais CamelCase
non_process
padronizar nomes de atributos do backend padronizar nomes de atributos do backend não utilizar mais camelcase
0
17,746
23,659,317,936
IssuesEvent
2022-08-26 14:10:30
saibrotech/mentoria
https://api.github.com/repos/saibrotech/mentoria
closed
Fazer processo seletivo JaBrasil TI
processo seletivo
https://jabrasil.org.br/ti Santa Catarina Aulas: Aulas online e ao vivo. Exercícios e conteúdos na plataforma Coursera. Você precisará se dedicar 5 vezes por semana, 4 horas por dia. Início das aulas: 29.08 a 09.12 Número de Vagas: 20 - [x] Fazer inscrição - [x] Enviar documentação necessária - [x] Participar dos encontros via Meet nos dias 22, 23 e 24 de agosto às 19h - [x] Compartilhar certificado da semana experimental no LinkedIn - [x] Aguardar resultado
1.0
Fazer processo seletivo JaBrasil TI - https://jabrasil.org.br/ti Santa Catarina Aulas: Aulas online e ao vivo. Exercícios e conteúdos na plataforma Coursera. Você precisará se dedicar 5 vezes por semana, 4 horas por dia. Início das aulas: 29.08 a 09.12 Número de Vagas: 20 - [x] Fazer inscrição - [x] Enviar documentação necessária - [x] Participar dos encontros via Meet nos dias 22, 23 e 24 de agosto às 19h - [x] Compartilhar certificado da semana experimental no LinkedIn - [x] Aguardar resultado
process
fazer processo seletivo jabrasil ti santa catarina aulas aulas online e ao vivo exercícios e conteúdos na plataforma coursera você precisará se dedicar vezes por semana horas por dia início das aulas a número de vagas fazer inscrição enviar documentação necessária participar dos encontros via meet nos dias e de agosto às compartilhar certificado da semana experimental no linkedin aguardar resultado
1
37,011
9,941,105,857
IssuesEvent
2019-07-03 10:44:05
ShaikASK/Testing
https://api.github.com/repos/ShaikASK/Testing
closed
Design Webform : Custom Webform : User is able to Create more then one custom webforms with same name
Custom Webforms Defect Design Webform HR Admin Module P2 Release #4 Build # 3
Steps To Replicate : 1.Launch the URL 2.Sign in “HR Admin” user 3.Go to Design Webforms 4.Create a “Custom Webform “ with some name say “Test” & save it 5. Create another webform with same name “Test” and save it Experienced Behavior : Observed that user is able to Create more then one custom web forms with same name (Refer Screen Shot) Expected Behavior : Ensure that application should not allow to create custom web forms with same name ![Design Webform](https://user-images.githubusercontent.com/31955617/60087530-cfc29f00-975a-11e9-8f75-afe3a4b8e5aa.png)
1.0
Design Webform : Custom Webform : User is able to Create more then one custom webforms with same name - Steps To Replicate : 1.Launch the URL 2.Sign in “HR Admin” user 3.Go to Design Webforms 4.Create a “Custom Webform “ with some name say “Test” & save it 5. Create another webform with same name “Test” and save it Experienced Behavior : Observed that user is able to Create more then one custom web forms with same name (Refer Screen Shot) Expected Behavior : Ensure that application should not allow to create custom web forms with same name ![Design Webform](https://user-images.githubusercontent.com/31955617/60087530-cfc29f00-975a-11e9-8f75-afe3a4b8e5aa.png)
non_process
design webform custom webform user is able to create more then one custom webforms with same name steps to replicate launch the url sign in “hr admin” user go to design webforms create a “custom webform “ with some name say “test” save it create another webform with same name “test” and save it experienced behavior observed that user is able to create more then one custom web forms with same name refer screen shot expected behavior ensure that application should not allow to create custom web forms with same name
0
728,572
25,084,885,234
IssuesEvent
2022-11-07 22:44:13
yugabyte/yugabyte-db
https://api.github.com/repos/yugabyte/yugabyte-db
closed
[YSQL] Add yb_pg_select_distinct.sql based on select_distinct.sql
kind/bug area/ysql priority/medium status/awaiting-triage
Jira Link: [DB-4062](https://yugabyte.atlassian.net/browse/DB-4062) ### Description This issue adds yb_select_distinct.sql which is based on select_distinct.sql The newly added file would be used for future backports involving select_distinct.sql
1.0
[YSQL] Add yb_pg_select_distinct.sql based on select_distinct.sql - Jira Link: [DB-4062](https://yugabyte.atlassian.net/browse/DB-4062) ### Description This issue adds yb_select_distinct.sql which is based on select_distinct.sql The newly added file would be used for future backports involving select_distinct.sql
non_process
add yb pg select distinct sql based on select distinct sql jira link description this issue adds yb select distinct sql which is based on select distinct sql the newly added file would be used for future backports involving select distinct sql
0
446,701
12,876,982,403
IssuesEvent
2020-07-11 08:20:39
threefoldtech/js-sdk
https://api.github.com/repos/threefoldtech/js-sdk
opened
starting nginx through starting threebot fails on osx
priority_major type_bug
psutil problem: ttps://github.com/giampaolo/psutil/issues/1447#issuecomment-646294879 ``` File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/bin/threebot", line 11, in <module> load_entry_point('js-sdk', 'console_scripts', 'threebot')() File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/lib/python3.8/site-packages/click/core.py", line 829, in __call__ return self.main(*args, **kwargs) File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/lib/python3.8/site-packages/click/core.py", line 782, in main rv = self.invoke(ctx) File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/lib/python3.8/site-packages/click/core.py", line 1259, in invoke return _process_result(sub_ctx.command.invoke(sub_ctx)) File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/lib/python3.8/site-packages/click/core.py", line 1066, in invoke return ctx.invoke(self.callback, **ctx.params) File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/lib/python3.8/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) File "/Users/rob/workspace/github.com/threefoldtech/js-sdk/jumpscale/entry_points/threebot.py", line 124, in start j.servers.threebot.start_default(wait=True, local=local) File "/Users/rob/workspace/github.com/threefoldtech/js-sdk/jumpscale/servers/threebot/__init__.py", line 29, in start_default server.start(wait=wait) File "/Users/rob/workspace/github.com/threefoldtech/js-sdk/jumpscale/servers/threebot/threebot.py", line 593, in start if self.is_running(): File "/Users/rob/workspace/github.com/threefoldtech/js-sdk/jumpscale/servers/threebot/threebot.py", line 506, in is_running nginx_running = self.nginx.is_running() File "/Users/rob/workspace/github.com/threefoldtech/js-sdk/jumpscale/tools/nginx/nginxserver.py", line 54, in is_running return j.tools.startupcmd.get(f"nginx_{self.name}").is_running() File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/src/js-ng/jumpscale/tools/startupcmd/startupcmd.py", line 225, in is_running if self.process: File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/src/js-ng/jumpscale/tools/startupcmd/startupcmd.py", line 89, in process if self.pid: File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/src/js-ng/jumpscale/tools/startupcmd/startupcmd.py", line 70, in pid pids = j.sals.process.get_pids(f"startupcmd_{self.instance_name}") File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/src/js-ng/jumpscale/sals/process/__init__.py", line 379, in get_pids elif match_predicate(process_name, process.exe()): File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/lib/python3.8/site-packages/psutil/__init__.py", line 673, in exe exe = self._proc.exe() File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/lib/python3.8/site-packages/psutil/_psosx.py", line 334, in wrapper return fun(self, *args, **kwargs) File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/lib/python3.8/site-packages/psutil/_psosx.py", line 413, in exe return cext.proc_exe(self.pid) FileNotFoundError: [Errno 2] No such file or directory (originated from proc_pidpath()) ```
1.0
starting nginx through starting threebot fails on osx - psutil problem: ttps://github.com/giampaolo/psutil/issues/1447#issuecomment-646294879 ``` File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/bin/threebot", line 11, in <module> load_entry_point('js-sdk', 'console_scripts', 'threebot')() File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/lib/python3.8/site-packages/click/core.py", line 829, in __call__ return self.main(*args, **kwargs) File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/lib/python3.8/site-packages/click/core.py", line 782, in main rv = self.invoke(ctx) File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/lib/python3.8/site-packages/click/core.py", line 1259, in invoke return _process_result(sub_ctx.command.invoke(sub_ctx)) File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/lib/python3.8/site-packages/click/core.py", line 1066, in invoke return ctx.invoke(self.callback, **ctx.params) File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/lib/python3.8/site-packages/click/core.py", line 610, in invoke return callback(*args, **kwargs) File "/Users/rob/workspace/github.com/threefoldtech/js-sdk/jumpscale/entry_points/threebot.py", line 124, in start j.servers.threebot.start_default(wait=True, local=local) File "/Users/rob/workspace/github.com/threefoldtech/js-sdk/jumpscale/servers/threebot/__init__.py", line 29, in start_default server.start(wait=wait) File "/Users/rob/workspace/github.com/threefoldtech/js-sdk/jumpscale/servers/threebot/threebot.py", line 593, in start if self.is_running(): File "/Users/rob/workspace/github.com/threefoldtech/js-sdk/jumpscale/servers/threebot/threebot.py", line 506, in is_running nginx_running = self.nginx.is_running() File "/Users/rob/workspace/github.com/threefoldtech/js-sdk/jumpscale/tools/nginx/nginxserver.py", line 54, in is_running return j.tools.startupcmd.get(f"nginx_{self.name}").is_running() File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/src/js-ng/jumpscale/tools/startupcmd/startupcmd.py", line 225, in is_running if self.process: File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/src/js-ng/jumpscale/tools/startupcmd/startupcmd.py", line 89, in process if self.pid: File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/src/js-ng/jumpscale/tools/startupcmd/startupcmd.py", line 70, in pid pids = j.sals.process.get_pids(f"startupcmd_{self.instance_name}") File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/src/js-ng/jumpscale/sals/process/__init__.py", line 379, in get_pids elif match_predicate(process_name, process.exe()): File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/lib/python3.8/site-packages/psutil/__init__.py", line 673, in exe exe = self._proc.exe() File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/lib/python3.8/site-packages/psutil/_psosx.py", line 334, in wrapper return fun(self, *args, **kwargs) File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/lib/python3.8/site-packages/psutil/_psosx.py", line 413, in exe return cext.proc_exe(self.pid) FileNotFoundError: [Errno 2] No such file or directory (originated from proc_pidpath()) ```
non_process
starting nginx through starting threebot fails on osx psutil problem ttps github com giampaolo psutil issues issuecomment file users rob library caches pypoetry virtualenvs js sdk bin threebot line in load entry point js sdk console scripts threebot file users rob library caches pypoetry virtualenvs js sdk lib site packages click core py line in call return self main args kwargs file users rob library caches pypoetry virtualenvs js sdk lib site packages click core py line in main rv self invoke ctx file users rob library caches pypoetry virtualenvs js sdk lib site packages click core py line in invoke return process result sub ctx command invoke sub ctx file users rob library caches pypoetry virtualenvs js sdk lib site packages click core py line in invoke return ctx invoke self callback ctx params file users rob library caches pypoetry virtualenvs js sdk lib site packages click core py line in invoke return callback args kwargs file users rob workspace github com threefoldtech js sdk jumpscale entry points threebot py line in start j servers threebot start default wait true local local file users rob workspace github com threefoldtech js sdk jumpscale servers threebot init py line in start default server start wait wait file users rob workspace github com threefoldtech js sdk jumpscale servers threebot threebot py line in start if self is running file users rob workspace github com threefoldtech js sdk jumpscale servers threebot threebot py line in is running nginx running self nginx is running file users rob workspace github com threefoldtech js sdk jumpscale tools nginx nginxserver py line in is running return j tools startupcmd get f nginx self name is running file users rob library caches pypoetry virtualenvs js sdk src js ng jumpscale tools startupcmd startupcmd py line in is running if self process file users rob library caches pypoetry virtualenvs js sdk src js ng jumpscale tools startupcmd startupcmd py line in process if self pid file users rob library caches pypoetry virtualenvs js sdk src js ng jumpscale tools startupcmd startupcmd py line in pid pids j sals process get pids f startupcmd self instance name file users rob library caches pypoetry virtualenvs js sdk src js ng jumpscale sals process init py line in get pids elif match predicate process name process exe file users rob library caches pypoetry virtualenvs js sdk lib site packages psutil init py line in exe exe self proc exe file users rob library caches pypoetry virtualenvs js sdk lib site packages psutil psosx py line in wrapper return fun self args kwargs file users rob library caches pypoetry virtualenvs js sdk lib site packages psutil psosx py line in exe return cext proc exe self pid filenotfounderror no such file or directory originated from proc pidpath
0
66,023
12,703,479,346
IssuesEvent
2020-06-22 22:25:33
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
opened
Attribute and analyzer for "soft abstract"
api-ready-for-review code-analyzer code-fixer
For abstract base types, we sometimes find ourselves in a situation where we'd like to add a new core concept to a type. In order to stay ABI-compatible, we can't add new `abstract` members as this means derived types that were compiled against the earlier version are missing the override, thus causing a `TypeLoadException`. The general pattern for this that we make the new members `virtual`, ideally with a workable implementation. This doesn't always work, so sometimes the only option is adding an implementation that throws `NotImplementedException`. A good example of this problem is in https://github.com/dotnet/runtime/issues/31362#issuecomment-643438552. A major downside of this approach is that developers deriving from the base type don't necessarily know that they are supposed to override these new virtual methods to get the best behavior. One option to solve this problem is by adding a new attribute and analyzer that warns when deriving from it without overriding the marked members. ### API Proposal ```C# namespace System.Diagnostics.CodeAnalysis { [AttributeUsage(AttributeTargets.Event | AttributeTargets.Method | AttributeTargets.Property)] public sealed class ShouldBeOverridenAttribute : Attribute { public ShouldBeOverridenAttribute(); } } ``` ### Analyzer Proposal Let's take [this example from the BCL](https://github.com/dotnet/runtime/issues/31362#issuecomment-643438552) and assume we marked the new virtual members as `[ShouldBeOverriden]` like this: ```C# namespace System.Threading.Channels { public abstract partial class ChannelReader<T> { [ShouldBeOverriden] public virtual bool CanPeek => false; [ShouldBeOverriden] public virtual bool TryPeek([MaybeNullWhen(false)] out T item) { item = default!; return false; } } } ``` A customer deriving from `ChannelReader<T>` would only override the `abstract` members by default, resulting in this: ```C# public class MyChannelReader<T> : ChannelReader<T> ~~~~~~~~~~~~~~~ { public override bool TryRead([MaybeNullWhen(false)] out T item) { throw new NotImplementedException(); } public override ValueTask<bool> WaitToReadAsync(CancellationToken cancellationToken = default) { throw new NotImplementedException(); } } ``` This would product warnings on the type delaration: ```text WARNING: ChannelReader<T>.CanPeek should be overriden WARNING: ChannelReader<T>.TryPeek(out T item) should be overriden ``` ### IDE We may want to change the behavior of the IDE so that the code fixer that generates overrides for abstract members automatically also generates overrides for any members marked as `[ShouldBeOverriden]`. /cc @dotnet/fxdc @CyrusNajmabadi
2.0
Attribute and analyzer for "soft abstract" - For abstract base types, we sometimes find ourselves in a situation where we'd like to add a new core concept to a type. In order to stay ABI-compatible, we can't add new `abstract` members as this means derived types that were compiled against the earlier version are missing the override, thus causing a `TypeLoadException`. The general pattern for this that we make the new members `virtual`, ideally with a workable implementation. This doesn't always work, so sometimes the only option is adding an implementation that throws `NotImplementedException`. A good example of this problem is in https://github.com/dotnet/runtime/issues/31362#issuecomment-643438552. A major downside of this approach is that developers deriving from the base type don't necessarily know that they are supposed to override these new virtual methods to get the best behavior. One option to solve this problem is by adding a new attribute and analyzer that warns when deriving from it without overriding the marked members. ### API Proposal ```C# namespace System.Diagnostics.CodeAnalysis { [AttributeUsage(AttributeTargets.Event | AttributeTargets.Method | AttributeTargets.Property)] public sealed class ShouldBeOverridenAttribute : Attribute { public ShouldBeOverridenAttribute(); } } ``` ### Analyzer Proposal Let's take [this example from the BCL](https://github.com/dotnet/runtime/issues/31362#issuecomment-643438552) and assume we marked the new virtual members as `[ShouldBeOverriden]` like this: ```C# namespace System.Threading.Channels { public abstract partial class ChannelReader<T> { [ShouldBeOverriden] public virtual bool CanPeek => false; [ShouldBeOverriden] public virtual bool TryPeek([MaybeNullWhen(false)] out T item) { item = default!; return false; } } } ``` A customer deriving from `ChannelReader<T>` would only override the `abstract` members by default, resulting in this: ```C# public class MyChannelReader<T> : ChannelReader<T> ~~~~~~~~~~~~~~~ { public override bool TryRead([MaybeNullWhen(false)] out T item) { throw new NotImplementedException(); } public override ValueTask<bool> WaitToReadAsync(CancellationToken cancellationToken = default) { throw new NotImplementedException(); } } ``` This would product warnings on the type delaration: ```text WARNING: ChannelReader<T>.CanPeek should be overriden WARNING: ChannelReader<T>.TryPeek(out T item) should be overriden ``` ### IDE We may want to change the behavior of the IDE so that the code fixer that generates overrides for abstract members automatically also generates overrides for any members marked as `[ShouldBeOverriden]`. /cc @dotnet/fxdc @CyrusNajmabadi
non_process
attribute and analyzer for soft abstract for abstract base types we sometimes find ourselves in a situation where we d like to add a new core concept to a type in order to stay abi compatible we can t add new abstract members as this means derived types that were compiled against the earlier version are missing the override thus causing a typeloadexception the general pattern for this that we make the new members virtual ideally with a workable implementation this doesn t always work so sometimes the only option is adding an implementation that throws notimplementedexception a good example of this problem is in a major downside of this approach is that developers deriving from the base type don t necessarily know that they are supposed to override these new virtual methods to get the best behavior one option to solve this problem is by adding a new attribute and analyzer that warns when deriving from it without overriding the marked members api proposal c namespace system diagnostics codeanalysis attributeusage attributetargets event attributetargets method attributetargets property public sealed class shouldbeoverridenattribute attribute public shouldbeoverridenattribute analyzer proposal let s take and assume we marked the new virtual members as like this c namespace system threading channels public abstract partial class channelreader public virtual bool canpeek false public virtual bool trypeek out t item item default return false a customer deriving from channelreader would only override the abstract members by default resulting in this c public class mychannelreader channelreader public override bool tryread out t item throw new notimplementedexception public override valuetask waittoreadasync cancellationtoken cancellationtoken default throw new notimplementedexception this would product warnings on the type delaration text warning channelreader canpeek should be overriden warning channelreader trypeek out t item should be overriden ide we may want to change the behavior of the ide so that the code fixer that generates overrides for abstract members automatically also generates overrides for any members marked as cc dotnet fxdc cyrusnajmabadi
0
442,803
12,750,625,150
IssuesEvent
2020-06-27 05:50:44
VandyHacks/vaken
https://api.github.com/repos/VandyHacks/vaken
opened
Add Summer Hackathon Waiver to the VH Summer 2020 Vaken Portal
Priority: High bite size good first issue
Add the summer hackathon waiver to application in the `event/vh2020-summer` branch.
1.0
Add Summer Hackathon Waiver to the VH Summer 2020 Vaken Portal - Add the summer hackathon waiver to application in the `event/vh2020-summer` branch.
non_process
add summer hackathon waiver to the vh summer vaken portal add the summer hackathon waiver to application in the event summer branch
0
127,762
17,360,351,711
IssuesEvent
2021-07-29 19:39:42
CMSgov/eAPD
https://api.github.com/repos/CMSgov/eAPD
opened
[Design Issue] Design a pattern for the export view for when non-required fields are not filled out
design
Some of the fields in the eAPD will not be required to be answered before submit. We have language that we use, noted in the content guide in the wiki, to note when a field isn't filled out so that it's clear to the reviewer that it wasn't a mistake or something broken in the system. Determine if we need any more language patterns or visual patterns to show when fields are not answered. ### This task is done when… - [ ] any acceptance criteria (not process oriented, requirements of feature) - [ ] designs are created, taking into consideration multiple FFY, multiple entries, long entries, weird entries, and/or the export view if applicable, validation considereations - [ ] 508 considerations and/or recommendations are reviewed - [ ] designs are shared - [ ] designs are approved by product - [ ] if there is a long discussion, a summary of changes for dev will be added to the last comment before transfer - [ ] a list of changes to be made if this is improving an existing feature --- Add additional labels (design, dev, compliance, BUG, etc) and size before submitting. If the issue is needed to complete prioritized work for the CURRENT SPRINT, add it to the "This Sprint" pipeline. Otherwise, all other issues will be automatically added to the _unprioritized pipeline_ for prioritization in backlog refinement or sprint planning with Product.
1.0
[Design Issue] Design a pattern for the export view for when non-required fields are not filled out - Some of the fields in the eAPD will not be required to be answered before submit. We have language that we use, noted in the content guide in the wiki, to note when a field isn't filled out so that it's clear to the reviewer that it wasn't a mistake or something broken in the system. Determine if we need any more language patterns or visual patterns to show when fields are not answered. ### This task is done when… - [ ] any acceptance criteria (not process oriented, requirements of feature) - [ ] designs are created, taking into consideration multiple FFY, multiple entries, long entries, weird entries, and/or the export view if applicable, validation considereations - [ ] 508 considerations and/or recommendations are reviewed - [ ] designs are shared - [ ] designs are approved by product - [ ] if there is a long discussion, a summary of changes for dev will be added to the last comment before transfer - [ ] a list of changes to be made if this is improving an existing feature --- Add additional labels (design, dev, compliance, BUG, etc) and size before submitting. If the issue is needed to complete prioritized work for the CURRENT SPRINT, add it to the "This Sprint" pipeline. Otherwise, all other issues will be automatically added to the _unprioritized pipeline_ for prioritization in backlog refinement or sprint planning with Product.
non_process
design a pattern for the export view for when non required fields are not filled out some of the fields in the eapd will not be required to be answered before submit we have language that we use noted in the content guide in the wiki to note when a field isn t filled out so that it s clear to the reviewer that it wasn t a mistake or something broken in the system determine if we need any more language patterns or visual patterns to show when fields are not answered this task is done when… any acceptance criteria not process oriented requirements of feature designs are created taking into consideration multiple ffy multiple entries long entries weird entries and or the export view if applicable validation considereations considerations and or recommendations are reviewed designs are shared designs are approved by product if there is a long discussion a summary of changes for dev will be added to the last comment before transfer a list of changes to be made if this is improving an existing feature add additional labels design dev compliance bug etc and size before submitting if the issue is needed to complete prioritized work for the current sprint add it to the this sprint pipeline otherwise all other issues will be automatically added to the unprioritized pipeline for prioritization in backlog refinement or sprint planning with product
0
45,749
5,730,838,307
IssuesEvent
2017-04-21 10:31:01
missioncommand/emp3-android
https://api.github.com/repos/missioncommand/emp3-android
closed
Test Tactial Graphics
integration testing Test Capability
Test drawing editing and renderering of the following MilStd tactical graphics in development. - [ ] Bridge => WAR.GRDTRK.EQT.GRDVEH.ENGVEH.BRG - [ ] bridge => TACTICAL GRAPHICS / MOBILITY/SURVIVABILITY / OBSTACLE BYPASS / CROSSING SITE/WATER CROSSING / BRIDGE OR GAP - [ ] Attack by Fire => TACTICAL GRAPHICS / COMMAND AND CONTROL AND GENERAL MANEUVER / OFFENSE / AREAS / ATTACK BY FIRE POSITION - [ ] Support by fire position => TACTICAL GRAPHICS / COMMAND AND CONTROL AND GENERAL MANEUVER / OFFENSE / AREAS / SUPPORT BY FIRE POSITION
2.0
Test Tactial Graphics - Test drawing editing and renderering of the following MilStd tactical graphics in development. - [ ] Bridge => WAR.GRDTRK.EQT.GRDVEH.ENGVEH.BRG - [ ] bridge => TACTICAL GRAPHICS / MOBILITY/SURVIVABILITY / OBSTACLE BYPASS / CROSSING SITE/WATER CROSSING / BRIDGE OR GAP - [ ] Attack by Fire => TACTICAL GRAPHICS / COMMAND AND CONTROL AND GENERAL MANEUVER / OFFENSE / AREAS / ATTACK BY FIRE POSITION - [ ] Support by fire position => TACTICAL GRAPHICS / COMMAND AND CONTROL AND GENERAL MANEUVER / OFFENSE / AREAS / SUPPORT BY FIRE POSITION
non_process
test tactial graphics test drawing editing and renderering of the following milstd tactical graphics in development bridge war grdtrk eqt grdveh engveh brg bridge tactical graphics mobility survivability obstacle bypass crossing site water crossing bridge or gap attack by fire tactical graphics command and control and general maneuver offense areas attack by fire position support by fire position tactical graphics command and control and general maneuver offense areas support by fire position
0
10,258
13,110,947,551
IssuesEvent
2020-08-04 21:45:16
SCIInstitute/Seg3D
https://api.github.com/repos/SCIInstitute/Seg3D
opened
Investigate VTK to replace OpenGL
graphics renderer software processes
Seg3D viewer features: Opacity Slice viewing Isosurfacing Brightness - gamma functions Colormaps Real-time feedback for labels Texture mapping Heads-up display Interactive feedback for paint, polylines, and other seed selection Volume rendering Optional: 3D interaction - seed point picking, moving, painting
1.0
Investigate VTK to replace OpenGL - Seg3D viewer features: Opacity Slice viewing Isosurfacing Brightness - gamma functions Colormaps Real-time feedback for labels Texture mapping Heads-up display Interactive feedback for paint, polylines, and other seed selection Volume rendering Optional: 3D interaction - seed point picking, moving, painting
process
investigate vtk to replace opengl viewer features opacity slice viewing isosurfacing brightness gamma functions colormaps real time feedback for labels texture mapping heads up display interactive feedback for paint polylines and other seed selection volume rendering optional interaction seed point picking moving painting
1
168,856
6,388,170,740
IssuesEvent
2017-08-03 15:04:20
fossasia/susi_skill_cms
https://api.github.com/repos/fossasia/susi_skill_cms
closed
Fix Dropdown menu structure and views across apps and pages
bug Priority: High
Related to: https://github.com/fossasia/chat.susi.ai/issues/657 **Not logged in users:** About -> Overview Chat -> /chat app Skills -> / skills app Settings -> /settings Login -> login **Logged in users:** About -> Overview Chat -> /chat app Skills -> / skills app Settings -> /settings [Themes -> theme popup (only in chat app)] Logout -> logout * Add icons to all menu items * Don't show "signup" on menu. The link to signup on the login page is sufficient. * On chat app direct settings to chat app settings * On skills cms direct settings link to skills settings * Ensure the dropdown menu works on the same way on chat pages overview etc.
1.0
Fix Dropdown menu structure and views across apps and pages - Related to: https://github.com/fossasia/chat.susi.ai/issues/657 **Not logged in users:** About -> Overview Chat -> /chat app Skills -> / skills app Settings -> /settings Login -> login **Logged in users:** About -> Overview Chat -> /chat app Skills -> / skills app Settings -> /settings [Themes -> theme popup (only in chat app)] Logout -> logout * Add icons to all menu items * Don't show "signup" on menu. The link to signup on the login page is sufficient. * On chat app direct settings to chat app settings * On skills cms direct settings link to skills settings * Ensure the dropdown menu works on the same way on chat pages overview etc.
non_process
fix dropdown menu structure and views across apps and pages related to not logged in users about overview chat chat app skills skills app settings settings login login logged in users about overview chat chat app skills skills app settings settings logout logout add icons to all menu items don t show signup on menu the link to signup on the login page is sufficient on chat app direct settings to chat app settings on skills cms direct settings link to skills settings ensure the dropdown menu works on the same way on chat pages overview etc
0
192,495
6,850,544,086
IssuesEvent
2017-11-14 03:55:54
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
hacks.mozilla.org - see bug description
browser-firefox priority-important
<!-- @browser: Firefox 57.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:57.0) Gecko/20100101 Firefox/57.0 --> <!-- @reported_with: web --> **URL**: https://hacks.mozilla.org/ **Browser / Version**: Firefox 57.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes **Problem type**: Something else **Description**: When I type the letter Y in the search bar it looks like a V. **Steps to Reproduce**: _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
hacks.mozilla.org - see bug description - <!-- @browser: Firefox 57.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:57.0) Gecko/20100101 Firefox/57.0 --> <!-- @reported_with: web --> **URL**: https://hacks.mozilla.org/ **Browser / Version**: Firefox 57.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes **Problem type**: Something else **Description**: When I type the letter Y in the search bar it looks like a V. **Steps to Reproduce**: _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
hacks mozilla org see bug description url browser version firefox operating system windows tested another browser yes problem type something else description when i type the letter y in the search bar it looks like a v steps to reproduce from with ❤️
0
20,056
26,542,945,083
IssuesEvent
2023-01-19 20:56:13
CSE201-project/PaperFriend-desktop-app
https://api.github.com/repos/CSE201-project/PaperFriend-desktop-app
closed
write an adapting welcome msg everytime the user log in
enhancement file processing
this msg will be different depending on how long the user was log out
1.0
write an adapting welcome msg everytime the user log in - this msg will be different depending on how long the user was log out
process
write an adapting welcome msg everytime the user log in this msg will be different depending on how long the user was log out
1
11,016
13,803,061,544
IssuesEvent
2020-10-11 00:27:07
SpencerTSterling/RatHutWebsite
https://api.github.com/repos/SpencerTSterling/RatHutWebsite
closed
Set up Continuous Integration with GitHub Actions
development process
GitHub actions should be used to build the project on each commit
1.0
Set up Continuous Integration with GitHub Actions - GitHub actions should be used to build the project on each commit
process
set up continuous integration with github actions github actions should be used to build the project on each commit
1
1,276
3,811,709,600
IssuesEvent
2016-03-27 01:23:08
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
opened
ServiceControllerTests.ControlCapabilities failed in CI on Win7
System.ServiceProcess
http://dotnet-ci.cloudapp.net/job/dotnet_corefx/job/outerloop_win7_debug/136/consoleFull ``` 09:29:26 System.ServiceProcess.Tests.ServiceControllerTests.ControlCapabilities [FAIL] 09:29:26 Assert.True() Failure 09:29:26 Expected: True 09:29:26 Actual: False 09:29:26 Stack Trace: 09:29:26 at System.ServiceProcess.Tests.ServiceControllerTests.ControlCapabilities() 09:29:27 System.Threading.Overlapped -> d:\j\workspace\outerloop_win---9c9e7d59\bin\ref\System.Th ```
1.0
ServiceControllerTests.ControlCapabilities failed in CI on Win7 - http://dotnet-ci.cloudapp.net/job/dotnet_corefx/job/outerloop_win7_debug/136/consoleFull ``` 09:29:26 System.ServiceProcess.Tests.ServiceControllerTests.ControlCapabilities [FAIL] 09:29:26 Assert.True() Failure 09:29:26 Expected: True 09:29:26 Actual: False 09:29:26 Stack Trace: 09:29:26 at System.ServiceProcess.Tests.ServiceControllerTests.ControlCapabilities() 09:29:27 System.Threading.Overlapped -> d:\j\workspace\outerloop_win---9c9e7d59\bin\ref\System.Th ```
process
servicecontrollertests controlcapabilities failed in ci on system serviceprocess tests servicecontrollertests controlcapabilities assert true failure expected true actual false stack trace at system serviceprocess tests servicecontrollertests controlcapabilities system threading overlapped d j workspace outerloop win bin ref system th
1
5,574
8,409,145,075
IssuesEvent
2018-10-12 06:00:12
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
Add support for security_opt: - "credentialspec=file://gmsaDocker.json"
area/agent area/windows kind/bug process/cherry-pick version/1.6
**Rancher Versions:** Server: 1.5.1 **Docker Version:** 17.03.0-ce **OS and where are the hosts located? (cloud, bare metal, etc):** windows server 2016 datacenter on hyper-v vm with nested virtualization enabled **Setup Details: (single node rancher vs. HA rancher, internal DB vs. external DB)** 2 rancher manager nodes, currently only 1 is in use, haproxy, 3 mariadb-s with galera **Environment Type: (Cattle/Kubernetes/Swarm/Mesos)** windows **Steps to Reproduce:** 1) create gmsa Active Directory account for the container https://docs.microsoft.com/en-us/virtualization/windowscontainers/manage-containers/manage-serviceaccounts 2) export credentials to json file 3) run windows container from rancher UI(or from rancher-cli) with docker-compose.yml **Results:** No success. If it will be done from the windows node(where rancher agent is installed) using docker CLI it'll be successful. Reason - that yml file with configuration of the rancher service has a string **credentialspec=file://gmsaDocker.json**, but this file will not be transmitted to server, even if is located in the same folder as the rancher-cli executable version: '2' services: test-express-23-dc: image: "microsoft/windowsservercore" network_mode: "transparent" security_opt: **- "credentialspec=file://gmsaDocker.json"** does not work because file is not transmitted to docker service That parameter works with docker cli only if there is a file gmsaDocker.json that is located in c:\ProgramData\Docker\Credentialspec **Suggested Improvements:** some way to provide this file from rancher-cli to rancher-server and to rancher-agent may be manually put that file to some location on the target machine, for example c:\programmdata\docker
1.0
Add support for security_opt: - "credentialspec=file://gmsaDocker.json" - **Rancher Versions:** Server: 1.5.1 **Docker Version:** 17.03.0-ce **OS and where are the hosts located? (cloud, bare metal, etc):** windows server 2016 datacenter on hyper-v vm with nested virtualization enabled **Setup Details: (single node rancher vs. HA rancher, internal DB vs. external DB)** 2 rancher manager nodes, currently only 1 is in use, haproxy, 3 mariadb-s with galera **Environment Type: (Cattle/Kubernetes/Swarm/Mesos)** windows **Steps to Reproduce:** 1) create gmsa Active Directory account for the container https://docs.microsoft.com/en-us/virtualization/windowscontainers/manage-containers/manage-serviceaccounts 2) export credentials to json file 3) run windows container from rancher UI(or from rancher-cli) with docker-compose.yml **Results:** No success. If it will be done from the windows node(where rancher agent is installed) using docker CLI it'll be successful. Reason - that yml file with configuration of the rancher service has a string **credentialspec=file://gmsaDocker.json**, but this file will not be transmitted to server, even if is located in the same folder as the rancher-cli executable version: '2' services: test-express-23-dc: image: "microsoft/windowsservercore" network_mode: "transparent" security_opt: **- "credentialspec=file://gmsaDocker.json"** does not work because file is not transmitted to docker service That parameter works with docker cli only if there is a file gmsaDocker.json that is located in c:\ProgramData\Docker\Credentialspec **Suggested Improvements:** some way to provide this file from rancher-cli to rancher-server and to rancher-agent may be manually put that file to some location on the target machine, for example c:\programmdata\docker
process
add support for security opt credentialspec file gmsadocker json rancher versions server docker version ce os and where are the hosts located cloud bare metal etc windows server datacenter on hyper v vm with nested virtualization enabled setup details single node rancher vs ha rancher internal db vs external db rancher manager nodes currently only is in use haproxy mariadb s with galera environment type cattle kubernetes swarm mesos windows steps to reproduce create gmsa active directory account for the container export credentials to json file run windows container from rancher ui or from rancher cli with docker compose yml results no success if it will be done from the windows node where rancher agent is installed using docker cli it ll be successful reason that yml file with configuration of the rancher service has a string credentialspec file gmsadocker json but this file will not be transmitted to server even if is located in the same folder as the rancher cli executable version services test express dc image microsoft windowsservercore network mode transparent security opt credentialspec file gmsadocker json does not work because file is not transmitted to docker service that parameter works with docker cli only if there is a file gmsadocker json that is located in c programdata docker credentialspec suggested improvements some way to provide this file from rancher cli to rancher server and to rancher agent may be manually put that file to some location on the target machine for example c programmdata docker
1
822,275
30,862,425,166
IssuesEvent
2023-08-03 05:01:38
WordPress/openverse
https://api.github.com/repos/WordPress/openverse
opened
Component: `VTag`
good first issue help wanted 🟨 priority: medium 🌟 goal: addition 🕹 aspect: interface 🧱 stack: frontend
## Description <!-- Describe the component, including different states. Do not include screenshots. --> The new `VTag` component will be used for clickable tags. The current, non-clickable, `VMediaTag` component will be replaced with it. To create a new component, run `just frontend/run create:component VTag`. This will run the component generator and will create the new component file, a Storybook story, unit and Playwright tests. The unit and playwright tests can be deleted. ## API <!-- Tentatively specify the props, state and emitted events of the component. --> ### Props <!-- Describe possible props and their types; or delete the section if the component takes no props. --> | Name | Type | Description | |-|-|-| | title | string | The tag text | | href | string | The url the tag should link to | ## Code samples <!-- Share pseudocode templates or high-level implementation code; or delete the section entirely. --> ``` <VButton as="VLink" size="small" variant="filled-gray" class="label-bold" :href="href" >{{ title }}</VButton> ``` <!-- Place a link to the Figma node of the component from the Design Library: https://www.figma.com/file/GIIQ4sDbaToCfFQyKMvzr8/Openverse-Design-Library --> - **Figma:** https://www.figma.com/file/niWnCgB7K0Y4e4mgxMrnRC/Additional-search-views?type=design&node-id=1263-62874&mode=design&t=WRJSFZNUbmpYERx8-4
1.0
Component: `VTag` - ## Description <!-- Describe the component, including different states. Do not include screenshots. --> The new `VTag` component will be used for clickable tags. The current, non-clickable, `VMediaTag` component will be replaced with it. To create a new component, run `just frontend/run create:component VTag`. This will run the component generator and will create the new component file, a Storybook story, unit and Playwright tests. The unit and playwright tests can be deleted. ## API <!-- Tentatively specify the props, state and emitted events of the component. --> ### Props <!-- Describe possible props and their types; or delete the section if the component takes no props. --> | Name | Type | Description | |-|-|-| | title | string | The tag text | | href | string | The url the tag should link to | ## Code samples <!-- Share pseudocode templates or high-level implementation code; or delete the section entirely. --> ``` <VButton as="VLink" size="small" variant="filled-gray" class="label-bold" :href="href" >{{ title }}</VButton> ``` <!-- Place a link to the Figma node of the component from the Design Library: https://www.figma.com/file/GIIQ4sDbaToCfFQyKMvzr8/Openverse-Design-Library --> - **Figma:** https://www.figma.com/file/niWnCgB7K0Y4e4mgxMrnRC/Additional-search-views?type=design&node-id=1263-62874&mode=design&t=WRJSFZNUbmpYERx8-4
non_process
component vtag description the new vtag component will be used for clickable tags the current non clickable vmediatag component will be replaced with it to create a new component run just frontend run create component vtag this will run the component generator and will create the new component file a storybook story unit and playwright tests the unit and playwright tests can be deleted api props name type description title string the tag text href string the url the tag should link to code samples vbutton as vlink size small variant filled gray class label bold href href title figma
0
576,457
17,087,317,763
IssuesEvent
2021-07-08 13:26:10
ParamagicDev/mrujs
https://api.github.com/repos/ParamagicDev/mrujs
opened
feat: Allow for end users to perform manual page caching
Priority: Normal Project Size: Small Type: Feature
#### Describe Problem Users cannot easily prefetch links #### Suggest Solution Expose the navigationAdapter's functions to fetch the cache.
1.0
feat: Allow for end users to perform manual page caching - #### Describe Problem Users cannot easily prefetch links #### Suggest Solution Expose the navigationAdapter's functions to fetch the cache.
non_process
feat allow for end users to perform manual page caching describe problem users cannot easily prefetch links suggest solution expose the navigationadapter s functions to fetch the cache
0
746,973
26,052,220,803
IssuesEvent
2022-12-22 20:01:11
evanshunt/derekstrap
https://api.github.com/repos/evanshunt/derekstrap
closed
1.0 ideas - Breakpoint method onBreakpoint();
enhancement Medium priority in progress
The `window.breakpointChanged` method can already be used to detect crossing a specific breakpoint threshold, but a simpler helper method that allowed you to pass in a breakpoint name and callbacks for what happens when the threshold is crossed would be handy.
1.0
1.0 ideas - Breakpoint method onBreakpoint(); - The `window.breakpointChanged` method can already be used to detect crossing a specific breakpoint threshold, but a simpler helper method that allowed you to pass in a breakpoint name and callbacks for what happens when the threshold is crossed would be handy.
non_process
ideas breakpoint method onbreakpoint the window breakpointchanged method can already be used to detect crossing a specific breakpoint threshold but a simpler helper method that allowed you to pass in a breakpoint name and callbacks for what happens when the threshold is crossed would be handy
0
41,446
8,973,322,044
IssuesEvent
2019-01-29 20:42:16
Foggalong/hardcode-fixer
https://api.github.com/repos/Foggalong/hardcode-fixer
closed
Return to castle wolfenstein SP
hardcoded app
| Application name | Icon name | Desktop file name | | --- | :-- | :-- | | Return to castle wolfenstein SP | `Icon=/usr/share/pixmaps/WolfSP.xpm` | `wolfsp.desktop` |
1.0
Return to castle wolfenstein SP - | Application name | Icon name | Desktop file name | | --- | :-- | :-- | | Return to castle wolfenstein SP | `Icon=/usr/share/pixmaps/WolfSP.xpm` | `wolfsp.desktop` |
non_process
return to castle wolfenstein sp application name icon name desktop file name return to castle wolfenstein sp icon usr share pixmaps wolfsp xpm wolfsp desktop
0
4,894
2,760,951,954
IssuesEvent
2015-04-28 14:56:16
mozilla/webmaker-app
https://api.github.com/repos/mozilla/webmaker-app
closed
UI - Bottom navigation components
design in progress
Finalize UI design for the bottom navigation components. Go to InVision to leave feedback: http://invis.io/BK2M18Q28
1.0
UI - Bottom navigation components - Finalize UI design for the bottom navigation components. Go to InVision to leave feedback: http://invis.io/BK2M18Q28
non_process
ui bottom navigation components finalize ui design for the bottom navigation components go to invision to leave feedback
0
173,981
27,554,871,758
IssuesEvent
2023-03-07 17:11:21
blockprotocol/blockprotocol
https://api.github.com/repos/blockprotocol/blockprotocol
opened
Set up visual regression testing with Playwright
area: design system
Look to leverage our existing BrowserStack subscription with something like BrowserStack Percy's integration w/ Playwright: https://docs.percy.io/docs/playwright
1.0
Set up visual regression testing with Playwright - Look to leverage our existing BrowserStack subscription with something like BrowserStack Percy's integration w/ Playwright: https://docs.percy.io/docs/playwright
non_process
set up visual regression testing with playwright look to leverage our existing browserstack subscription with something like browserstack percy s integration w playwright
0
10,931
13,745,297,207
IssuesEvent
2020-10-06 02:26:59
Alphadrake86/CraftsMadeByHand
https://api.github.com/repos/Alphadrake86/CraftsMadeByHand
closed
Set up CI pipeline
Dev Processes
Set up Continuous Integration pipeline to ensure all incoming commits compile
1.0
Set up CI pipeline - Set up Continuous Integration pipeline to ensure all incoming commits compile
process
set up ci pipeline set up continuous integration pipeline to ensure all incoming commits compile
1
49,263
13,186,563,223
IssuesEvent
2020-08-13 00:34:17
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
clsim - hobo clsim build chokes doc build (Trac #1077)
Incomplete Migration Migrated from Trac combo simulation defect
<details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1077">https://code.icecube.wisc.edu/ticket/1077</a>, reported by nega and owned by claudio.kopper</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:15:08", "description": "clsim's dependency list should probably AND'd, not OR'd. \n\n{{{\n-- + clsim\n-- +-- python [symlinks] \n-- +-- Geant4 or OpenCL is not installed on your system. clsim will fail if it is not used with parameterizations. \n-- +-- numpy support (for tabulator) \n-- +-- safeprimes_base32.gz already downloaded \n-- +-- gmp support (make_safeprimes utility) \n-- +-- clsim-pybindings \n}}}\n\ngives:\n{{{\n[ 62%] Generating html from icetray-inspect of clsim\nIgnoring 'clsim': dlopen() dynamic loading error: /home/nega/i3/combo/build/lib/libclsim.so: undefined symbol: clCreateSubDevices/home/nega/i3/combo/build/CMakeFiles/clsim-inspection.xml:4: parser error : Opening and ending tag mismatch: project line 3 and icetray-inspect\n</icetray-inspect>\n ^\n/home/nega/i3/combo/build/CMakeFiles/clsim-inspection.xml:5: parser error : Premature end of data in tag icetray-inspect line 2\n\n^\nunable to parse /home/nega/i3/combo/build/CMakeFiles/clsim-inspection.xml\nmake[3]: *** [clsim/CMakeFiles/clsim-clsim-inspect] Error 6\nmake[2]: *** [clsim/CMakeFiles/clsim-clsim-inspect.dir/all] Error 2\nmake[1]: *** [CMakeFiles/docs.dir/rule] Error 2\nmake: *** [docs] Error 2\n \n~/i3/combo/build 1m 53s\n\u276f \n}}}", "reporter": "nega", "cc": "", "resolution": "worksforme", "_ts": "1550067308113782", "component": "combo simulation", "summary": "clsim - hobo clsim build chokes doc build", "priority": "normal", "keywords": "clsim docs", "time": "2015-07-30T04:18:17", "milestone": "", "owner": "claudio.kopper", "type": "defect" } ``` </p> </details>
1.0
clsim - hobo clsim build chokes doc build (Trac #1077) - <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1077">https://code.icecube.wisc.edu/ticket/1077</a>, reported by nega and owned by claudio.kopper</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:15:08", "description": "clsim's dependency list should probably AND'd, not OR'd. \n\n{{{\n-- + clsim\n-- +-- python [symlinks] \n-- +-- Geant4 or OpenCL is not installed on your system. clsim will fail if it is not used with parameterizations. \n-- +-- numpy support (for tabulator) \n-- +-- safeprimes_base32.gz already downloaded \n-- +-- gmp support (make_safeprimes utility) \n-- +-- clsim-pybindings \n}}}\n\ngives:\n{{{\n[ 62%] Generating html from icetray-inspect of clsim\nIgnoring 'clsim': dlopen() dynamic loading error: /home/nega/i3/combo/build/lib/libclsim.so: undefined symbol: clCreateSubDevices/home/nega/i3/combo/build/CMakeFiles/clsim-inspection.xml:4: parser error : Opening and ending tag mismatch: project line 3 and icetray-inspect\n</icetray-inspect>\n ^\n/home/nega/i3/combo/build/CMakeFiles/clsim-inspection.xml:5: parser error : Premature end of data in tag icetray-inspect line 2\n\n^\nunable to parse /home/nega/i3/combo/build/CMakeFiles/clsim-inspection.xml\nmake[3]: *** [clsim/CMakeFiles/clsim-clsim-inspect] Error 6\nmake[2]: *** [clsim/CMakeFiles/clsim-clsim-inspect.dir/all] Error 2\nmake[1]: *** [CMakeFiles/docs.dir/rule] Error 2\nmake: *** [docs] Error 2\n \n~/i3/combo/build 1m 53s\n\u276f \n}}}", "reporter": "nega", "cc": "", "resolution": "worksforme", "_ts": "1550067308113782", "component": "combo simulation", "summary": "clsim - hobo clsim build chokes doc build", "priority": "normal", "keywords": "clsim docs", "time": "2015-07-30T04:18:17", "milestone": "", "owner": "claudio.kopper", "type": "defect" } ``` </p> </details>
non_process
clsim hobo clsim build chokes doc build trac migrated from json status closed changetime description clsim s dependency list should probably and d not or d n n n clsim n python n or opencl is not installed on your system clsim will fail if it is not used with parameterizations n numpy support for tabulator n safeprimes gz already downloaded n gmp support make safeprimes utility n clsim pybindings n n ngives n n generating html from icetray inspect of clsim nignoring clsim dlopen dynamic loading error home nega combo build lib libclsim so undefined symbol clcreatesubdevices home nega combo build cmakefiles clsim inspection xml parser error opening and ending tag mismatch project line and icetray inspect n n n home nega combo build cmakefiles clsim inspection xml parser error premature end of data in tag icetray inspect line n n nunable to parse home nega combo build cmakefiles clsim inspection xml nmake error nmake error nmake error nmake error n n combo build n n reporter nega cc resolution worksforme ts component combo simulation summary clsim hobo clsim build chokes doc build priority normal keywords clsim docs time milestone owner claudio kopper type defect
0
18,978
24,966,042,743
IssuesEvent
2022-11-01 19:28:28
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
qgis_process temporary files are not cleaned up
Processing Bug
### What is the bug or the crash? On each call to `qgis_process`, two new files are created in the user's temporary directory, but they are not removed afterwards: - an empty file `QGIS3.xxxxxx` - a file in a subdirectory: `QGIS3-yyyyyy/zzzzzz_styles.db`. This is a SQLite database with 14 tables having 0 rows. The xxx/yyy/zzz strings seem random. With several tested algorithms, these files all appear to remain empty / with 0 rows. If this is always the case, then perhaps the files shouldn't be there? When checking the R package qgisprocess, this behaviour yields a note about remaining detritus (https://github.com/paleolimbot/qgisprocess/issues/62). ### Steps to reproduce the issue In a Linux shell: ```bash $ find /tmp -type f \( -name "QGIS*" -o -name "*_styles.db" \) $ $ qgis_process &> /dev/null $ $ find /tmp -type f \( -name "QGIS*" -o -name "*_styles.db" \) /tmp/QGIS3-UJyKwM/KHMJUP_styles.db /tmp/QGIS3.BHULds $ $ qgis_process &> /dev/null $ $ find /tmp -type f \( -name "QGIS*" -o -name "*_styles.db" \) /tmp/QGIS3-UJyKwM/KHMJUP_styles.db /tmp/QGIS3.BHULds /tmp/QGIS3.YBTuhF /tmp/QGIS3-Ovuiht/hHUlfq_styles.db ``` ### Versions <!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.0//EN" "http://www.w3.org/TR/REC-html40/strict.dtd"> <html><head><meta http-equiv="Content-Type" content="text/html; charset=utf-8" /></head><body> QGIS version | 3.26.3-Buenos Aires | QGIS code revision | 65e4edfdad -- | -- | -- | -- Qt version | 5.12.8 Python version | 3.8.10 GDAL/OGR version | 3.4.3 PROJ version | 8.2.0 EPSG Registry database version | v10.038 (2021-10-21) GEOS version | 3.10.2-CAPI-1.16.0 SQLite version | 3.31.1 PDAL version | 2.2.0 PostgreSQL client version | 12.12 (Ubuntu 12.12-0ubuntu0.20.04.1) SpatiaLite version | 5.0.1 QWT version | 6.1.4 QScintilla2 version | 2.11.2 OS version | Linux Mint 20   |   |   |   Active Python plugins geopunt4Qgis | 2.2.4 ViewshedAnalysis | 1.7 cartography_tools | 1.2.1 quick_map_services | 0.19.27 grassprovider | 2.12.99 processing | 2.12.99 sagaprovider | 2.12.99 db_manager | 0.1.20 MetaSearch | 0.3.6 otbprovider | 2.12.99 </body></html> ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [ ] I tried with a new QGIS profile ### Additional context _No response_
1.0
qgis_process temporary files are not cleaned up - ### What is the bug or the crash? On each call to `qgis_process`, two new files are created in the user's temporary directory, but they are not removed afterwards: - an empty file `QGIS3.xxxxxx` - a file in a subdirectory: `QGIS3-yyyyyy/zzzzzz_styles.db`. This is a SQLite database with 14 tables having 0 rows. The xxx/yyy/zzz strings seem random. With several tested algorithms, these files all appear to remain empty / with 0 rows. If this is always the case, then perhaps the files shouldn't be there? When checking the R package qgisprocess, this behaviour yields a note about remaining detritus (https://github.com/paleolimbot/qgisprocess/issues/62). ### Steps to reproduce the issue In a Linux shell: ```bash $ find /tmp -type f \( -name "QGIS*" -o -name "*_styles.db" \) $ $ qgis_process &> /dev/null $ $ find /tmp -type f \( -name "QGIS*" -o -name "*_styles.db" \) /tmp/QGIS3-UJyKwM/KHMJUP_styles.db /tmp/QGIS3.BHULds $ $ qgis_process &> /dev/null $ $ find /tmp -type f \( -name "QGIS*" -o -name "*_styles.db" \) /tmp/QGIS3-UJyKwM/KHMJUP_styles.db /tmp/QGIS3.BHULds /tmp/QGIS3.YBTuhF /tmp/QGIS3-Ovuiht/hHUlfq_styles.db ``` ### Versions <!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.0//EN" "http://www.w3.org/TR/REC-html40/strict.dtd"> <html><head><meta http-equiv="Content-Type" content="text/html; charset=utf-8" /></head><body> QGIS version | 3.26.3-Buenos Aires | QGIS code revision | 65e4edfdad -- | -- | -- | -- Qt version | 5.12.8 Python version | 3.8.10 GDAL/OGR version | 3.4.3 PROJ version | 8.2.0 EPSG Registry database version | v10.038 (2021-10-21) GEOS version | 3.10.2-CAPI-1.16.0 SQLite version | 3.31.1 PDAL version | 2.2.0 PostgreSQL client version | 12.12 (Ubuntu 12.12-0ubuntu0.20.04.1) SpatiaLite version | 5.0.1 QWT version | 6.1.4 QScintilla2 version | 2.11.2 OS version | Linux Mint 20   |   |   |   Active Python plugins geopunt4Qgis | 2.2.4 ViewshedAnalysis | 1.7 cartography_tools | 1.2.1 quick_map_services | 0.19.27 grassprovider | 2.12.99 processing | 2.12.99 sagaprovider | 2.12.99 db_manager | 0.1.20 MetaSearch | 0.3.6 otbprovider | 2.12.99 </body></html> ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [ ] I tried with a new QGIS profile ### Additional context _No response_
process
qgis process temporary files are not cleaned up what is the bug or the crash on each call to qgis process two new files are created in the user s temporary directory but they are not removed afterwards an empty file xxxxxx a file in a subdirectory yyyyyy zzzzzz styles db this is a sqlite database with tables having rows the xxx yyy zzz strings seem random with several tested algorithms these files all appear to remain empty with rows if this is always the case then perhaps the files shouldn t be there when checking the r package qgisprocess this behaviour yields a note about remaining detritus steps to reproduce the issue in a linux shell bash find tmp type f name qgis o name styles db qgis process dev null find tmp type f name qgis o name styles db tmp ujykwm khmjup styles db tmp bhulds qgis process dev null find tmp type f name qgis o name styles db tmp ujykwm khmjup styles db tmp bhulds tmp ybtuhf tmp ovuiht hhulfq styles db versions doctype html public dtd html en qgis version buenos aires qgis code revision qt version python version gdal ogr version proj version epsg registry database version geos version capi sqlite version pdal version postgresql client version ubuntu spatialite version qwt version version os version linux mint         active python plugins viewshedanalysis cartography tools quick map services grassprovider processing sagaprovider db manager metasearch otbprovider supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context no response
1
699,358
24,013,874,252
IssuesEvent
2022-09-14 21:37:41
tl-its-umich-edu/my-learning-analytics
https://api.github.com/repos/tl-its-umich-edu/my-learning-analytics
closed
Vulnerability: CVE-2015-20107, Python
⚠️ priority ☠ vulnerability
From [vulnerabilities](https://docs.google.com/spreadsheets/d/12cyQTkIxCxwP0lAi58eFvznBcKDO3yB3mKdSenGkBOM) spreadsheet, based on Unizin analysis: CVE ID | Effective Severity | Severity | Impacted Image | Vulnerable Package | Remediated Package | URL -- | -- | -- | -- | -- | -- | -- CVE-2015-20107 | CRITICAL | CRITICAL | gcr.io/unizin-core/myla:2022.01.01 | python3.9/3.9.2-1 | python3.9/MAXIMUM | https://security-tracker.debian.org/tracker/CVE-2015-20107
1.0
Vulnerability: CVE-2015-20107, Python - From [vulnerabilities](https://docs.google.com/spreadsheets/d/12cyQTkIxCxwP0lAi58eFvznBcKDO3yB3mKdSenGkBOM) spreadsheet, based on Unizin analysis: CVE ID | Effective Severity | Severity | Impacted Image | Vulnerable Package | Remediated Package | URL -- | -- | -- | -- | -- | -- | -- CVE-2015-20107 | CRITICAL | CRITICAL | gcr.io/unizin-core/myla:2022.01.01 | python3.9/3.9.2-1 | python3.9/MAXIMUM | https://security-tracker.debian.org/tracker/CVE-2015-20107
non_process
vulnerability cve python from spreadsheet based on unizin analysis cve id effective severity severity impacted image vulnerable package remediated package url cve critical critical gcr io unizin core myla maximum
0
198,566
22,659,667,161
IssuesEvent
2022-07-02 01:15:21
snykiotcubedev/arangodb-3.7.6
https://api.github.com/repos/snykiotcubedev/arangodb-3.7.6
reopened
CVE-2021-23382 (Medium) detected in postcss-7.0.32.tgz
security vulnerability
## CVE-2021-23382 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>postcss-7.0.32.tgz</b></p></summary> <p>Tool for transforming styles with JS plugins</p> <p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-7.0.32.tgz">https://registry.npmjs.org/postcss/-/postcss-7.0.32.tgz</a></p> <p> Dependency Hierarchy: - css-loader-2.1.1.tgz (Root Library) - :x: **postcss-7.0.32.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/snykiotcubedev/arangodb-3.7.6/commit/fce8f85f1c2f070c8e6a8e76d17210a2117d3833">fce8f85f1c2f070c8e6a8e76d17210a2117d3833</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package postcss before 8.2.13 are vulnerable to Regular Expression Denial of Service (ReDoS) via getAnnotationURL() and loadAnnotation() in lib/previous-map.js. The vulnerable regexes are caused mainly by the sub-pattern \/\*\s* sourceMappingURL=(.*). <p>Publish Date: 2021-04-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23382>CVE-2021-23382</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23382">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23382</a></p> <p>Release Date: 2021-04-26</p> <p>Fix Resolution (postcss): 7.0.36</p> <p>Direct dependency fix Resolution (css-loader): 3.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-23382 (Medium) detected in postcss-7.0.32.tgz - ## CVE-2021-23382 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>postcss-7.0.32.tgz</b></p></summary> <p>Tool for transforming styles with JS plugins</p> <p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-7.0.32.tgz">https://registry.npmjs.org/postcss/-/postcss-7.0.32.tgz</a></p> <p> Dependency Hierarchy: - css-loader-2.1.1.tgz (Root Library) - :x: **postcss-7.0.32.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/snykiotcubedev/arangodb-3.7.6/commit/fce8f85f1c2f070c8e6a8e76d17210a2117d3833">fce8f85f1c2f070c8e6a8e76d17210a2117d3833</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package postcss before 8.2.13 are vulnerable to Regular Expression Denial of Service (ReDoS) via getAnnotationURL() and loadAnnotation() in lib/previous-map.js. The vulnerable regexes are caused mainly by the sub-pattern \/\*\s* sourceMappingURL=(.*). <p>Publish Date: 2021-04-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23382>CVE-2021-23382</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23382">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23382</a></p> <p>Release Date: 2021-04-26</p> <p>Fix Resolution (postcss): 7.0.36</p> <p>Direct dependency fix Resolution (css-loader): 3.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in postcss tgz cve medium severity vulnerability vulnerable library postcss tgz tool for transforming styles with js plugins library home page a href dependency hierarchy css loader tgz root library x postcss tgz vulnerable library found in head commit a href found in base branch main vulnerability details the package postcss before are vulnerable to regular expression denial of service redos via getannotationurl and loadannotation in lib previous map js the vulnerable regexes are caused mainly by the sub pattern s sourcemappingurl publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution postcss direct dependency fix resolution css loader step up your open source security game with mend
0
17,492
23,305,507,934
IssuesEvent
2022-08-07 23:50:04
lynnandtonic/nestflix.fun
https://api.github.com/repos/lynnandtonic/nestflix.fun
closed
Add Lil Ark Angel “Cats Don’t Dance” (Screenshots and Title Screen Added)
suggested title in process rental
Please add as much of the following info as you can: Title: Lil Ark Angel Type (film/tv show): Film - Golden Age musical Film or show in which it appears: Cats Don't Dance Is the parent film/show streaming anywhere? Yes - Amazon Prime About when in the parent film/show does it appear? probably about the first third of the film Actual footage of the film/show can be seen (yes/no)? Yes. Here is the musical number "Little Boat on the Sea": https://www.youtube.com/watch?v=bShn4m6I49k Cast: Darla Dimple Director: Flanagan Production Company: Mammoth Pictures Quote: I built a little boat as cute as it can be, and I put all the animals two-by-two in my little boat on the sea.
1.0
Add Lil Ark Angel “Cats Don’t Dance” (Screenshots and Title Screen Added) - Please add as much of the following info as you can: Title: Lil Ark Angel Type (film/tv show): Film - Golden Age musical Film or show in which it appears: Cats Don't Dance Is the parent film/show streaming anywhere? Yes - Amazon Prime About when in the parent film/show does it appear? probably about the first third of the film Actual footage of the film/show can be seen (yes/no)? Yes. Here is the musical number "Little Boat on the Sea": https://www.youtube.com/watch?v=bShn4m6I49k Cast: Darla Dimple Director: Flanagan Production Company: Mammoth Pictures Quote: I built a little boat as cute as it can be, and I put all the animals two-by-two in my little boat on the sea.
process
add lil ark angel “cats don’t dance” screenshots and title screen added please add as much of the following info as you can title lil ark angel type film tv show film golden age musical film or show in which it appears cats don t dance is the parent film show streaming anywhere yes amazon prime about when in the parent film show does it appear probably about the first third of the film actual footage of the film show can be seen yes no yes here is the musical number little boat on the sea cast darla dimple director flanagan production company mammoth pictures quote i built a little boat as cute as it can be and i put all the animals two by two in my little boat on the sea
1
11,413
14,242,136,981
IssuesEvent
2020-11-19 01:02:44
googleapis/java-speech
https://api.github.com/repos/googleapis/java-speech
closed
longRunningRecognizeAsync in asyncRecognizeGcs doesn't work with long audio files.
api: speech type: process
While troubleshooting this [issue](https://github.com/GoogleCloudPlatform/php-docs-samples/issues/1094) I discovered that the ```longRunningRecognizeAsync ```in this [sample](https://github.com/googleapis/java-speech/blob/master/samples/snippets/src/main/java/com/example/speech/Recognize.java) presents a similar behavior as the original issue I was looking into. To repro: Use the ```asyncRecognizeGcs ``` from the [samples](https://github.com/googleapis/java-speech/blob/master/samples/snippets/src/main/java/com/example/speech/Recognize.java) and feed it a long gcs audio file (at least 30 minutes long) and run it. The process will run for a while to render no results. Will see a "Task was cancelled." message.
1.0
longRunningRecognizeAsync in asyncRecognizeGcs doesn't work with long audio files. - While troubleshooting this [issue](https://github.com/GoogleCloudPlatform/php-docs-samples/issues/1094) I discovered that the ```longRunningRecognizeAsync ```in this [sample](https://github.com/googleapis/java-speech/blob/master/samples/snippets/src/main/java/com/example/speech/Recognize.java) presents a similar behavior as the original issue I was looking into. To repro: Use the ```asyncRecognizeGcs ``` from the [samples](https://github.com/googleapis/java-speech/blob/master/samples/snippets/src/main/java/com/example/speech/Recognize.java) and feed it a long gcs audio file (at least 30 minutes long) and run it. The process will run for a while to render no results. Will see a "Task was cancelled." message.
process
longrunningrecognizeasync in asyncrecognizegcs doesn t work with long audio files while troubleshooting this i discovered that the longrunningrecognizeasync in this presents a similar behavior as the original issue i was looking into to repro use the asyncrecognizegcs from the and feed it a long gcs audio file at least minutes long and run it the process will run for a while to render no results will see a task was cancelled message
1
9,858
12,857,494,627
IssuesEvent
2020-07-09 09:23:51
zammad/zammad
https://api.github.com/repos/zammad/zammad
opened
ServiceNow mails from other service providers are not detected
bug mail processing
### Infos: * Used Zammad version: 3.4 * Installation method (source, package, ..): all * Operating system: all * Database + version: all * Elasticsearch version: all * Browser + version: all ### Expected behavior: * ServiceNow mails should be detected based on the mail headers, instead of the sender address service-now.com ### Actual behavior: * ServiceNow mails from different sender addresses than service-now.com are not detected. ### Steps to reproduce the behavior: -
1.0
ServiceNow mails from other service providers are not detected - ### Infos: * Used Zammad version: 3.4 * Installation method (source, package, ..): all * Operating system: all * Database + version: all * Elasticsearch version: all * Browser + version: all ### Expected behavior: * ServiceNow mails should be detected based on the mail headers, instead of the sender address service-now.com ### Actual behavior: * ServiceNow mails from different sender addresses than service-now.com are not detected. ### Steps to reproduce the behavior: -
process
servicenow mails from other service providers are not detected infos used zammad version installation method source package all operating system all database version all elasticsearch version all browser version all expected behavior servicenow mails should be detected based on the mail headers instead of the sender address service now com actual behavior servicenow mails from different sender addresses than service now com are not detected steps to reproduce the behavior
1
11,972
14,737,010,373
IssuesEvent
2021-01-07 00:36:50
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
not able to make cc charges from billing site
anc-external anc-process anp-important anp-urgent ant-bug ant-support
In GitLab by @kdjstudios on Apr 9, 2018, 15:48 **Submitted by:** Gaylan Garrett <gaylan@keenercom.net> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-04-09-52796/conversation **Server:** Hosted **Client/Site:** Keener **Account:** NA **Issue:** I just wanted to make you aware that I am not able to do a credit card charge from the site. I have to go to authorize.net and make the charge and then return to SA billing to post the charge.
1.0
not able to make cc charges from billing site - In GitLab by @kdjstudios on Apr 9, 2018, 15:48 **Submitted by:** Gaylan Garrett <gaylan@keenercom.net> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-04-09-52796/conversation **Server:** Hosted **Client/Site:** Keener **Account:** NA **Issue:** I just wanted to make you aware that I am not able to do a credit card charge from the site. I have to go to authorize.net and make the charge and then return to SA billing to post the charge.
process
not able to make cc charges from billing site in gitlab by kdjstudios on apr submitted by gaylan garrett helpdesk server hosted client site keener account na issue i just wanted to make you aware that i am not able to do a credit card charge from the site i have to go to authorize net and make the charge and then return to sa billing to post the charge
1
10,318
13,160,837,615
IssuesEvent
2020-08-10 18:21:52
GoogleCloudPlatform/stackdriver-sandbox
https://api.github.com/repos/GoogleCloudPlatform/stackdriver-sandbox
closed
Move stackdriver-sandbox.dev webpage off of github master branch
priority: p2 type: process
Currently stackdriver-sandbox.dev points to the content of the docs folder in the master branch. Whenever we make a change there, it's reflected on the website immediately. This prevents us from making incremental improvements that should not be visible to external customers until the release happens. We should address that by pointing stackdriver-sandbox.dev to a hosted version of the website on AppEngine or somewhere else on GCP.
1.0
Move stackdriver-sandbox.dev webpage off of github master branch - Currently stackdriver-sandbox.dev points to the content of the docs folder in the master branch. Whenever we make a change there, it's reflected on the website immediately. This prevents us from making incremental improvements that should not be visible to external customers until the release happens. We should address that by pointing stackdriver-sandbox.dev to a hosted version of the website on AppEngine or somewhere else on GCP.
process
move stackdriver sandbox dev webpage off of github master branch currently stackdriver sandbox dev points to the content of the docs folder in the master branch whenever we make a change there it s reflected on the website immediately this prevents us from making incremental improvements that should not be visible to external customers until the release happens we should address that by pointing stackdriver sandbox dev to a hosted version of the website on appengine or somewhere else on gcp
1
5,634
8,485,285,383
IssuesEvent
2018-10-26 07:14:53
thewca/wca-regulations
https://api.github.com/repos/thewca/wca-regulations
opened
Temporary "100-week" events
announcement-worthy process proposal
I would like to seriously consider [Arnaud's suggestion of a temporary event](https://www.speedsolving.com/forum/threads/idea-alternating-1-year-event.70659/). He suggests 1 year, but I think we should go for ≈2 years so that every puzzle can have a world champion once. 2 years minus a month of gap sounds appropriate, so I'm calling these "100-week events" in my head. :-D We have had several discussions that touch on the issue that that adding or removing events is a big deal. Major competitions have a lot of pressure to hold all events, and the combination of 1) modern scheduling and 2) the lack of a maintained database for unofficial events at official competitions means that unofficial events are very uncommon. Rough idea for an event that would be active from 2020 to 2021: - In January 2019 we announce some number (say, ≈3) of events that we want to consider for the next cycle. We encourage competitions to hold these unofficially in order to work out the format and give the competitors a chance to try it out. - Around August (after the major competitions are held), we hold a vote of all active WCA competitors for one event. - The event becomes an official event from January 2020 to the end of 2021. - In January 2021, we start the cycle for a new puzzle. We have already retired events in the past, so this shouldn't require special tweaking from the website. What I really like about this idea is that it lowers the commitment for adding an official event, without setting unfair expectations and fueling debates about which events are worth keeping. It also prevents keeping around an event for so long that I believe that new events are a benefit to the community because: - We regularly get questions or requests for new events, so competitors want them! Listening to community wishes may become more important as we try to stay relevant compared to other efforts like Red Bull. - If we regularly engage with the community for this one point of feedback, we have a chance to keep community conversations going between WCA officials and competitors. (I feel like we have a tendency for the two to disconnect over time.) - The cubing community is large enough that puzzle manufacturers are likely to invest resources into puzzles that we consider for 100-week events. And once a well-engineered puzzle is on the market, it will hopefully stay available for longer. I would love for us to incentivize perpetual innovation in new types of speed puzzles. - Novel puzzles encourage research into algorithms and software development, whose effects also stay around. e now see a lot of cubers focused on advanced techniques and tweaks to existing methods, but I think we're losing out on more classical development of methods. - I would really like to see us explore less classic event formats, such as team BLD or FMC duel.
1.0
Temporary "100-week" events - I would like to seriously consider [Arnaud's suggestion of a temporary event](https://www.speedsolving.com/forum/threads/idea-alternating-1-year-event.70659/). He suggests 1 year, but I think we should go for ≈2 years so that every puzzle can have a world champion once. 2 years minus a month of gap sounds appropriate, so I'm calling these "100-week events" in my head. :-D We have had several discussions that touch on the issue that that adding or removing events is a big deal. Major competitions have a lot of pressure to hold all events, and the combination of 1) modern scheduling and 2) the lack of a maintained database for unofficial events at official competitions means that unofficial events are very uncommon. Rough idea for an event that would be active from 2020 to 2021: - In January 2019 we announce some number (say, ≈3) of events that we want to consider for the next cycle. We encourage competitions to hold these unofficially in order to work out the format and give the competitors a chance to try it out. - Around August (after the major competitions are held), we hold a vote of all active WCA competitors for one event. - The event becomes an official event from January 2020 to the end of 2021. - In January 2021, we start the cycle for a new puzzle. We have already retired events in the past, so this shouldn't require special tweaking from the website. What I really like about this idea is that it lowers the commitment for adding an official event, without setting unfair expectations and fueling debates about which events are worth keeping. It also prevents keeping around an event for so long that I believe that new events are a benefit to the community because: - We regularly get questions or requests for new events, so competitors want them! Listening to community wishes may become more important as we try to stay relevant compared to other efforts like Red Bull. - If we regularly engage with the community for this one point of feedback, we have a chance to keep community conversations going between WCA officials and competitors. (I feel like we have a tendency for the two to disconnect over time.) - The cubing community is large enough that puzzle manufacturers are likely to invest resources into puzzles that we consider for 100-week events. And once a well-engineered puzzle is on the market, it will hopefully stay available for longer. I would love for us to incentivize perpetual innovation in new types of speed puzzles. - Novel puzzles encourage research into algorithms and software development, whose effects also stay around. e now see a lot of cubers focused on advanced techniques and tweaks to existing methods, but I think we're losing out on more classical development of methods. - I would really like to see us explore less classic event formats, such as team BLD or FMC duel.
process
temporary week events i would like to seriously consider he suggests year but i think we should go for ≈ years so that every puzzle can have a world champion once years minus a month of gap sounds appropriate so i m calling these week events in my head d we have had several discussions that touch on the issue that that adding or removing events is a big deal major competitions have a lot of pressure to hold all events and the combination of modern scheduling and the lack of a maintained database for unofficial events at official competitions means that unofficial events are very uncommon rough idea for an event that would be active from to in january we announce some number say ≈ of events that we want to consider for the next cycle we encourage competitions to hold these unofficially in order to work out the format and give the competitors a chance to try it out around august after the major competitions are held we hold a vote of all active wca competitors for one event the event becomes an official event from january to the end of in january we start the cycle for a new puzzle we have already retired events in the past so this shouldn t require special tweaking from the website what i really like about this idea is that it lowers the commitment for adding an official event without setting unfair expectations and fueling debates about which events are worth keeping it also prevents keeping around an event for so long that i believe that new events are a benefit to the community because we regularly get questions or requests for new events so competitors want them listening to community wishes may become more important as we try to stay relevant compared to other efforts like red bull if we regularly engage with the community for this one point of feedback we have a chance to keep community conversations going between wca officials and competitors i feel like we have a tendency for the two to disconnect over time the cubing community is large enough that puzzle manufacturers are likely to invest resources into puzzles that we consider for week events and once a well engineered puzzle is on the market it will hopefully stay available for longer i would love for us to incentivize perpetual innovation in new types of speed puzzles novel puzzles encourage research into algorithms and software development whose effects also stay around e now see a lot of cubers focused on advanced techniques and tweaks to existing methods but i think we re losing out on more classical development of methods i would really like to see us explore less classic event formats such as team bld or fmc duel
1
10,169
8,403,543,181
IssuesEvent
2018-10-11 10:03:42
nest/nest-simulator
https://api.github.com/repos/nest/nest-simulator
closed
Index of user contributed modules/extensions (MyModule)
C: Infrastructure I: No breaking change P: In progess S: Low T: Maintenance
As far as I can see it there is currently no index for user contributed modules hosted on the nest website or the github wiki (https://github.com/nest/nest-simulator/wiki). As more specialized modules are contributed I think it would make sense to maintain a list (possibly sorted by use-case) to point the community to available modules.
1.0
Index of user contributed modules/extensions (MyModule) - As far as I can see it there is currently no index for user contributed modules hosted on the nest website or the github wiki (https://github.com/nest/nest-simulator/wiki). As more specialized modules are contributed I think it would make sense to maintain a list (possibly sorted by use-case) to point the community to available modules.
non_process
index of user contributed modules extensions mymodule as far as i can see it there is currently no index for user contributed modules hosted on the nest website or the github wiki as more specialized modules are contributed i think it would make sense to maintain a list possibly sorted by use case to point the community to available modules
0
5,268
8,058,213,508
IssuesEvent
2018-08-02 17:45:41
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
[DOTX031E][ERROR] The file 'topic.dita' is not available to resolve link information in XHTML transformation when using chunk='to-content' feature
obsolete preprocess/chunking
The XHTML transformation fails when a DITA map uses the `chunk='to-content'` feature and there are related-links to the merged topics. The error that I get is: ``` [topicpull] Recoverable error on line 576 of topicpullImpl.xsl: [topicpull] FODC0002: java.io.FileNotFoundException: [topicpull] C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml\topics\flowers\concepts\springFlowers.dita (The system cannot find the path specified) [topicpull] file:/C:/Users/radu_pisoi/Desktop/temp/dita-issue/issue2/flowers/topics/flowers/snowdrop.dita:9:33: [DOTX031E][ERROR]: The file file:/C:/Users/radu_pisoi/Desktop/temp/dita-issue/issue2/flowers/temp/xhtml/topics/flowers/concepts/springFlowers.dita is not available to resolve link information. ``` I noticed that if I remove the related-link to the topic merged in the parent topic, then the issue disappears. ``` Executing: "D:\workspace\eXml\jre\jre1.8.0_72-x64/bin/java" "-Dhttp.proxySet=true" "-Dhttp.proxyHost=proxy.sync.ro" "-Dhttps.proxyHost=proxy.sync.ro" "-Dhttp.proxyPort=3128" "-Dhttps.proxyPort=3128" -Xmx384m "-Doxygen.org.apache.xerces.xni.parser.XMLParserConfiguration=org.ditang.relaxng.defaults.RelaxDefaultsParserConfiguration" -classpath "D:\workspace\eXml\tools\ant/lib/ant-launcher.jar" "-Dant.home=D:\workspace\eXml\tools\ant" org.apache.tools.ant.launch.Launcher -lib "D:\workspace\dita-ot-2.3.1\lib\saxon-9.1.0.8-dom.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\saxon-9.1.0.8.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\commons-io-2.4.jar" -lib "D:\workspace\eXml\classes" -lib "D:\workspace\eXml\lib\oxygen-content-completion-api.jar" -lib "D:\workspace\eXml\lib\oxygen-css-pretty-printer.jar" -lib "D:\workspace\eXml\lib\oxygen-css-validator.jar" -lib "D:\workspace\eXml\lib\oxygen-emf.jar" -lib "D:\workspace\eXml\lib\oxygen-jfx-components.jar" -lib "D:\workspace\eXml\lib\oxygen-text-search.jar" -lib "D:\workspace\eXml\lib\oxygen-token-markers.jar" -lib "D:\workspace\eXml\lib\oxygen-validation-api.jar" -lib "D:\workspace\eXml\lib\oxygen-xquery-pretty-printer.jar" -lib "D:\workspace\eXml\lib\resolver.jar" -lib "D:\workspace\eXml\lib\oxygen-token-markers.jar" -lib "D:\workspace\eXml\lib\org.eclipse.wst.xml.xpath2.processor_1.2.0.jar" -lib "D:\workspace\eXml\lib\xml-apis.jar" -lib "D:\workspace\eXml\lib\xercesImpl.jar" -lib "D:\workspace\eXml\lib\commons-io-1.3.1.jar" -lib "D:\workspace\eXml\lib\commons-logging-1.2.jar" -lib "D:\workspace\eXml\lib\log4j.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\commons-codec-1.9.jar" -lib "D:\workspace\eXml\lib\jing.jar" -lib "D:\workspace\eXml\lib\saxon9ee.jar" -lib "D:\workspace\eXml\lib\saxon.jar" -lib "D:\workspace\eXml\lib\xmlgraphics-commons-2.1.jar" -lib "D:\workspace\eXml\lib\fop.jar" -lib "D:\workspace\eXml\lib\batik-all-1.8.jar" -lib "D:\workspace\eXml\lib\js.jar" -lib "D:\workspace\eXml\lib\poi-3.10-FINAL-20140208.jar" -lib "D:\workspace\eXml\lib\nekohtml.jar" -lib "D:\workspace\eXml\lib\xml-apis-ext.jar" -lib "D:\workspace\eXml\lib\avalon-framework-api-4.3.1.jar" -lib "D:\workspace\eXml\lib\avalon-framework-impl-4.3.1.jar" -lib "D:\workspace\eXml\lib\jeuclid-core.jar" -lib "D:\workspace\eXml\lib\jeuclid-fop.jar" -lib "D:\workspace\eXml\lib\jai_tiff.jar" -lib "D:\workspace\eXml\lib\jh.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\jsearch.jar" -lib "D:\workspace\eXml\lib\lucene-analyzers-common-6.1.0.jar" -lib "D:\workspace\eXml\lib\lucene-analyzers-kuromoji-6.1.0.jar" -lib "D:\workspace\eXml\lib\lucene-core-6.1.0.jar" -lib "D:\workspace\eXml\lib\lucene-misc-6.1.0.jar" -lib "D:\workspace\eXml\lib\lucene-queries-6.1.0.jar" -lib "D:\workspace\eXml\lib\lucene-queryparser-6.1.0.jar" -lib "D:\workspace\eXml\lib\lucene-suggest-6.1.0.jar" -lib "D:\workspace\dita-ot-2.3.1" -lib "D:\workspace\dita-ot-2.3.1\lib\dost-patches.jar" -lib "D:\workspace\dita-ot-2.3.1\lib" -lib "D:\workspace\dita-ot-2.3.1\lib\ant-apache-resolver-1.9.4.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\ant-launcher.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\ant.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\commons-codec-1.9.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\commons-io-2.4.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\dost-configuration.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\dost.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\guava-19.0.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\icu4j-54.1.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\saxon-9.1.0.8-dom.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\saxon-9.1.0.8.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\xercesImpl-2.11.0.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\xml-apis-1.4.01.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\xml-resolver-1.2.jar" -lib "D:\workspace\dita-ot-2.3.1\plugins\org.dita.odt\lib\odt.jar" -lib "D:\workspace\dita-ot-2.3.1\plugins\org.dita.pdf2\lib\fo.jar" -lib "D:\workspace\dita-ot-2.3.1\plugins\org.dita.pdf2.axf\lib\axf.jar" -lib "D:\workspace\dita-ot-2.3.1\plugins\org.dita.pdf2.xep\lib\xep.jar" -lib "D:\workspace\dita-ot-2.3.1\plugins\org.dita.wordrtf\lib\wordrtf.jar" -f "D:\workspace\dita-ot-2.3.1/build.xml" "-Dtranstype=xhtml" "-Dbasedir=C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers" "-Doutput.dir=C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers/out/xhtml" "-Ddita.temp.dir=C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers/temp/xhtml" "-Dargs.input=C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\flowers.ditamap" "-Ddita.dir=D:\workspace\dita-ot-2.3.1" "-DbaseJVMArgLine=-Xmx384m" Buildfile: D:\workspace\dita-ot-2.3.1\build.xml init: dita2xhtml.init: check-arg: [mkdir] Created dir: C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\out\xhtml [mkdir] Created dir: C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml log-arg: [echo] ***************************************************************** [echo] * basedir = C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers [echo] * dita.dir = D:\workspace\dita-ot-2.3.1 [echo] * transtype = xhtml [echo] * tempdir = C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers/temp/xhtml [echo] * outputdir = C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers/out/xhtml [echo] * clean.temp = true [echo] * DITA-OT version = 2.3.1 [echo] * XML parser = Xerces [echo] * XSLT processor = Saxon [echo] * collator = ICU [echo] ***************************************************************** [echo] #Ant properties [echo] #Fri Aug 19 17:37:23 EEST 2016 [echo] args.grammar.cache=yes [echo] args.input=C\:\\Users\\radu_pisoi\\Desktop\\temp\\dita-issue\\issue2\\flowers\\flowers.ditamap [echo] args.logdir=C\:\\Users\\radu_pisoi\\Desktop\\temp\\dita-issue\\issue2\\flowers/out/xhtml [echo] args.xml.systemid.set=yes [echo] dita.dir=D\:\\workspace\\dita-ot-2.3.1 [echo] dita.plugin.com.sophos.tocjs.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\com.sophos.tocjs [echo] dita.plugin.org.dita.base.dir=D\:\\workspace\\dita-ot-2.3.1 [echo] dita.plugin.org.dita.docbook.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.docbook [echo] dita.plugin.org.dita.eclipsecontent.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.eclipsecontent [echo] dita.plugin.org.dita.eclipsehelp.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.eclipsehelp [echo] dita.plugin.org.dita.html5.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.html5 [echo] dita.plugin.org.dita.htmlhelp.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.htmlhelp [echo] dita.plugin.org.dita.javahelp.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.javahelp [echo] dita.plugin.org.dita.odt.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.odt [echo] dita.plugin.org.dita.pdf2.axf.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.pdf2.axf [echo] dita.plugin.org.dita.pdf2.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.pdf2 [echo] dita.plugin.org.dita.pdf2.fop.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.pdf2.fop [echo] dita.plugin.org.dita.pdf2.xep.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.pdf2.xep [echo] dita.plugin.org.dita.specialization.dita11.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.specialization.dita11 [echo] dita.plugin.org.dita.specialization.eclipsemap.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.specialization.eclipsemap [echo] dita.plugin.org.dita.troff.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.troff [echo] dita.plugin.org.dita.wordrtf.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.wordrtf [echo] dita.plugin.org.dita.xhtml.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.xhtml [echo] dita.plugin.org.oasis-open.dita.v1_2.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.oasis-open.dita.v1_2 [echo] dita.plugin.org.oasis-open.dita.v1_3.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.oasis-open.dita.v1_3 [echo] dita.temp.dir=C\:\\Users\\radu_pisoi\\Desktop\\temp\\dita-issue\\issue2\\flowers/temp/xhtml [echo] ***************************************************************** -------------------------------------------------- topicpull-check: topicpull: [topicpull] Transforming into C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml [topicpull] Loading stylesheet D:\workspace\dita-ot-2.3.1\xsl\preprocess\topicpull.xsl [topicpull] Processing C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml\topics\index.dita [topicpull] Processing C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml\concepts\springFlowers.dita [topicpull] Recoverable error on line 576 of topicpullImpl.xsl: [topicpull] FODC0002: java.io.FileNotFoundException: [topicpull] C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml\topics\flowers\concepts\springFlowers.dita (The system cannot find the path specified) [topicpull] file:/C:/Users/radu_pisoi/Desktop/temp/dita-issue/issue2/flowers/topics/flowers/snowdrop.dita:9:33: [DOTX031E][ERROR]: The file file:/C:/Users/radu_pisoi/Desktop/temp/dita-issue/issue2/flowers/temp/xhtml/topics/flowers/concepts/springFlowers.dita is not available to resolve link information. [topicpull] Recoverable error on line 654 of topicpullImpl.xsl: [topicpull] FODC0002: java.io.FileNotFoundException: [topicpull] C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml\topics\flowers\concepts\springFlowers.dita (The system cannot find the path specified) [topicpull] Recoverable error on line 1108 of topicpullImpl.xsl: [topicpull] FODC0002: java.io.FileNotFoundException: [topicpull] C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml\topics\flowers\concepts\springFlowers.dita (The system cannot find the path specified) [topicpull] Recoverable error on line 1844 of topicpullImpl.xsl: [topicpull] FODC0002: java.io.FileNotFoundException: [topicpull] C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml\topics\flowers\concepts\springFlowers.dita (The system cannot find the path specified) ``` I've tested with DITA-OT 2.3.1 and the attached DITA map: [flowers.zip](https://github.com/dita-ot/dita-ot/files/427257/flowers.zip)
1.0
[DOTX031E][ERROR] The file 'topic.dita' is not available to resolve link information in XHTML transformation when using chunk='to-content' feature - The XHTML transformation fails when a DITA map uses the `chunk='to-content'` feature and there are related-links to the merged topics. The error that I get is: ``` [topicpull] Recoverable error on line 576 of topicpullImpl.xsl: [topicpull] FODC0002: java.io.FileNotFoundException: [topicpull] C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml\topics\flowers\concepts\springFlowers.dita (The system cannot find the path specified) [topicpull] file:/C:/Users/radu_pisoi/Desktop/temp/dita-issue/issue2/flowers/topics/flowers/snowdrop.dita:9:33: [DOTX031E][ERROR]: The file file:/C:/Users/radu_pisoi/Desktop/temp/dita-issue/issue2/flowers/temp/xhtml/topics/flowers/concepts/springFlowers.dita is not available to resolve link information. ``` I noticed that if I remove the related-link to the topic merged in the parent topic, then the issue disappears. ``` Executing: "D:\workspace\eXml\jre\jre1.8.0_72-x64/bin/java" "-Dhttp.proxySet=true" "-Dhttp.proxyHost=proxy.sync.ro" "-Dhttps.proxyHost=proxy.sync.ro" "-Dhttp.proxyPort=3128" "-Dhttps.proxyPort=3128" -Xmx384m "-Doxygen.org.apache.xerces.xni.parser.XMLParserConfiguration=org.ditang.relaxng.defaults.RelaxDefaultsParserConfiguration" -classpath "D:\workspace\eXml\tools\ant/lib/ant-launcher.jar" "-Dant.home=D:\workspace\eXml\tools\ant" org.apache.tools.ant.launch.Launcher -lib "D:\workspace\dita-ot-2.3.1\lib\saxon-9.1.0.8-dom.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\saxon-9.1.0.8.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\commons-io-2.4.jar" -lib "D:\workspace\eXml\classes" -lib "D:\workspace\eXml\lib\oxygen-content-completion-api.jar" -lib "D:\workspace\eXml\lib\oxygen-css-pretty-printer.jar" -lib "D:\workspace\eXml\lib\oxygen-css-validator.jar" -lib "D:\workspace\eXml\lib\oxygen-emf.jar" -lib "D:\workspace\eXml\lib\oxygen-jfx-components.jar" -lib "D:\workspace\eXml\lib\oxygen-text-search.jar" -lib "D:\workspace\eXml\lib\oxygen-token-markers.jar" -lib "D:\workspace\eXml\lib\oxygen-validation-api.jar" -lib "D:\workspace\eXml\lib\oxygen-xquery-pretty-printer.jar" -lib "D:\workspace\eXml\lib\resolver.jar" -lib "D:\workspace\eXml\lib\oxygen-token-markers.jar" -lib "D:\workspace\eXml\lib\org.eclipse.wst.xml.xpath2.processor_1.2.0.jar" -lib "D:\workspace\eXml\lib\xml-apis.jar" -lib "D:\workspace\eXml\lib\xercesImpl.jar" -lib "D:\workspace\eXml\lib\commons-io-1.3.1.jar" -lib "D:\workspace\eXml\lib\commons-logging-1.2.jar" -lib "D:\workspace\eXml\lib\log4j.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\commons-codec-1.9.jar" -lib "D:\workspace\eXml\lib\jing.jar" -lib "D:\workspace\eXml\lib\saxon9ee.jar" -lib "D:\workspace\eXml\lib\saxon.jar" -lib "D:\workspace\eXml\lib\xmlgraphics-commons-2.1.jar" -lib "D:\workspace\eXml\lib\fop.jar" -lib "D:\workspace\eXml\lib\batik-all-1.8.jar" -lib "D:\workspace\eXml\lib\js.jar" -lib "D:\workspace\eXml\lib\poi-3.10-FINAL-20140208.jar" -lib "D:\workspace\eXml\lib\nekohtml.jar" -lib "D:\workspace\eXml\lib\xml-apis-ext.jar" -lib "D:\workspace\eXml\lib\avalon-framework-api-4.3.1.jar" -lib "D:\workspace\eXml\lib\avalon-framework-impl-4.3.1.jar" -lib "D:\workspace\eXml\lib\jeuclid-core.jar" -lib "D:\workspace\eXml\lib\jeuclid-fop.jar" -lib "D:\workspace\eXml\lib\jai_tiff.jar" -lib "D:\workspace\eXml\lib\jh.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\jsearch.jar" -lib "D:\workspace\eXml\lib\lucene-analyzers-common-6.1.0.jar" -lib "D:\workspace\eXml\lib\lucene-analyzers-kuromoji-6.1.0.jar" -lib "D:\workspace\eXml\lib\lucene-core-6.1.0.jar" -lib "D:\workspace\eXml\lib\lucene-misc-6.1.0.jar" -lib "D:\workspace\eXml\lib\lucene-queries-6.1.0.jar" -lib "D:\workspace\eXml\lib\lucene-queryparser-6.1.0.jar" -lib "D:\workspace\eXml\lib\lucene-suggest-6.1.0.jar" -lib "D:\workspace\dita-ot-2.3.1" -lib "D:\workspace\dita-ot-2.3.1\lib\dost-patches.jar" -lib "D:\workspace\dita-ot-2.3.1\lib" -lib "D:\workspace\dita-ot-2.3.1\lib\ant-apache-resolver-1.9.4.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\ant-launcher.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\ant.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\commons-codec-1.9.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\commons-io-2.4.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\dost-configuration.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\dost.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\guava-19.0.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\icu4j-54.1.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\saxon-9.1.0.8-dom.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\saxon-9.1.0.8.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\xercesImpl-2.11.0.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\xml-apis-1.4.01.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\xml-resolver-1.2.jar" -lib "D:\workspace\dita-ot-2.3.1\plugins\org.dita.odt\lib\odt.jar" -lib "D:\workspace\dita-ot-2.3.1\plugins\org.dita.pdf2\lib\fo.jar" -lib "D:\workspace\dita-ot-2.3.1\plugins\org.dita.pdf2.axf\lib\axf.jar" -lib "D:\workspace\dita-ot-2.3.1\plugins\org.dita.pdf2.xep\lib\xep.jar" -lib "D:\workspace\dita-ot-2.3.1\plugins\org.dita.wordrtf\lib\wordrtf.jar" -f "D:\workspace\dita-ot-2.3.1/build.xml" "-Dtranstype=xhtml" "-Dbasedir=C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers" "-Doutput.dir=C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers/out/xhtml" "-Ddita.temp.dir=C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers/temp/xhtml" "-Dargs.input=C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\flowers.ditamap" "-Ddita.dir=D:\workspace\dita-ot-2.3.1" "-DbaseJVMArgLine=-Xmx384m" Buildfile: D:\workspace\dita-ot-2.3.1\build.xml init: dita2xhtml.init: check-arg: [mkdir] Created dir: C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\out\xhtml [mkdir] Created dir: C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml log-arg: [echo] ***************************************************************** [echo] * basedir = C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers [echo] * dita.dir = D:\workspace\dita-ot-2.3.1 [echo] * transtype = xhtml [echo] * tempdir = C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers/temp/xhtml [echo] * outputdir = C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers/out/xhtml [echo] * clean.temp = true [echo] * DITA-OT version = 2.3.1 [echo] * XML parser = Xerces [echo] * XSLT processor = Saxon [echo] * collator = ICU [echo] ***************************************************************** [echo] #Ant properties [echo] #Fri Aug 19 17:37:23 EEST 2016 [echo] args.grammar.cache=yes [echo] args.input=C\:\\Users\\radu_pisoi\\Desktop\\temp\\dita-issue\\issue2\\flowers\\flowers.ditamap [echo] args.logdir=C\:\\Users\\radu_pisoi\\Desktop\\temp\\dita-issue\\issue2\\flowers/out/xhtml [echo] args.xml.systemid.set=yes [echo] dita.dir=D\:\\workspace\\dita-ot-2.3.1 [echo] dita.plugin.com.sophos.tocjs.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\com.sophos.tocjs [echo] dita.plugin.org.dita.base.dir=D\:\\workspace\\dita-ot-2.3.1 [echo] dita.plugin.org.dita.docbook.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.docbook [echo] dita.plugin.org.dita.eclipsecontent.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.eclipsecontent [echo] dita.plugin.org.dita.eclipsehelp.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.eclipsehelp [echo] dita.plugin.org.dita.html5.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.html5 [echo] dita.plugin.org.dita.htmlhelp.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.htmlhelp [echo] dita.plugin.org.dita.javahelp.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.javahelp [echo] dita.plugin.org.dita.odt.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.odt [echo] dita.plugin.org.dita.pdf2.axf.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.pdf2.axf [echo] dita.plugin.org.dita.pdf2.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.pdf2 [echo] dita.plugin.org.dita.pdf2.fop.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.pdf2.fop [echo] dita.plugin.org.dita.pdf2.xep.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.pdf2.xep [echo] dita.plugin.org.dita.specialization.dita11.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.specialization.dita11 [echo] dita.plugin.org.dita.specialization.eclipsemap.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.specialization.eclipsemap [echo] dita.plugin.org.dita.troff.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.troff [echo] dita.plugin.org.dita.wordrtf.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.wordrtf [echo] dita.plugin.org.dita.xhtml.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.xhtml [echo] dita.plugin.org.oasis-open.dita.v1_2.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.oasis-open.dita.v1_2 [echo] dita.plugin.org.oasis-open.dita.v1_3.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.oasis-open.dita.v1_3 [echo] dita.temp.dir=C\:\\Users\\radu_pisoi\\Desktop\\temp\\dita-issue\\issue2\\flowers/temp/xhtml [echo] ***************************************************************** -------------------------------------------------- topicpull-check: topicpull: [topicpull] Transforming into C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml [topicpull] Loading stylesheet D:\workspace\dita-ot-2.3.1\xsl\preprocess\topicpull.xsl [topicpull] Processing C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml\topics\index.dita [topicpull] Processing C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml\concepts\springFlowers.dita [topicpull] Recoverable error on line 576 of topicpullImpl.xsl: [topicpull] FODC0002: java.io.FileNotFoundException: [topicpull] C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml\topics\flowers\concepts\springFlowers.dita (The system cannot find the path specified) [topicpull] file:/C:/Users/radu_pisoi/Desktop/temp/dita-issue/issue2/flowers/topics/flowers/snowdrop.dita:9:33: [DOTX031E][ERROR]: The file file:/C:/Users/radu_pisoi/Desktop/temp/dita-issue/issue2/flowers/temp/xhtml/topics/flowers/concepts/springFlowers.dita is not available to resolve link information. [topicpull] Recoverable error on line 654 of topicpullImpl.xsl: [topicpull] FODC0002: java.io.FileNotFoundException: [topicpull] C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml\topics\flowers\concepts\springFlowers.dita (The system cannot find the path specified) [topicpull] Recoverable error on line 1108 of topicpullImpl.xsl: [topicpull] FODC0002: java.io.FileNotFoundException: [topicpull] C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml\topics\flowers\concepts\springFlowers.dita (The system cannot find the path specified) [topicpull] Recoverable error on line 1844 of topicpullImpl.xsl: [topicpull] FODC0002: java.io.FileNotFoundException: [topicpull] C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml\topics\flowers\concepts\springFlowers.dita (The system cannot find the path specified) ``` I've tested with DITA-OT 2.3.1 and the attached DITA map: [flowers.zip](https://github.com/dita-ot/dita-ot/files/427257/flowers.zip)
process
the file topic dita is not available to resolve link information in xhtml transformation when using chunk to content feature the xhtml transformation fails when a dita map uses the chunk to content feature and there are related links to the merged topics the error that i get is recoverable error on line of topicpullimpl xsl java io filenotfoundexception c users radu pisoi desktop temp dita issue flowers temp xhtml topics flowers concepts springflowers dita the system cannot find the path specified file c users radu pisoi desktop temp dita issue flowers topics flowers snowdrop dita the file file c users radu pisoi desktop temp dita issue flowers temp xhtml topics flowers concepts springflowers dita is not available to resolve link information i noticed that if i remove the related link to the topic merged in the parent topic then the issue disappears executing d workspace exml jre bin java dhttp proxyset true dhttp proxyhost proxy sync ro dhttps proxyhost proxy sync ro dhttp proxyport dhttps proxyport doxygen org apache xerces xni parser xmlparserconfiguration org ditang relaxng defaults relaxdefaultsparserconfiguration classpath d workspace exml tools ant lib ant launcher jar dant home d workspace exml tools ant org apache tools ant launch launcher lib d workspace dita ot lib saxon dom jar lib d workspace dita ot lib saxon jar lib d workspace dita ot lib commons io jar lib d workspace exml classes lib d workspace exml lib oxygen content completion api jar lib d workspace exml lib oxygen css pretty printer jar lib d workspace exml lib oxygen css validator jar lib d workspace exml lib oxygen emf jar lib d workspace exml lib oxygen jfx components jar lib d workspace exml lib oxygen text search jar lib d workspace exml lib oxygen token markers jar lib d workspace exml lib oxygen validation api jar lib d workspace exml lib oxygen xquery pretty printer jar lib d workspace exml lib resolver jar lib d workspace exml lib oxygen token markers jar lib d workspace exml lib org eclipse wst xml processor jar lib d workspace exml lib xml apis jar lib d workspace exml lib xercesimpl jar lib d workspace exml lib commons io jar lib d workspace exml lib commons logging jar lib d workspace exml lib jar lib d workspace dita ot lib commons codec jar lib d workspace exml lib jing jar lib d workspace exml lib jar lib d workspace exml lib saxon jar lib d workspace exml lib xmlgraphics commons jar lib d workspace exml lib fop jar lib d workspace exml lib batik all jar lib d workspace exml lib js jar lib d workspace exml lib poi final jar lib d workspace exml lib nekohtml jar lib d workspace exml lib xml apis ext jar lib d workspace exml lib avalon framework api jar lib d workspace exml lib avalon framework impl jar lib d workspace exml lib jeuclid core jar lib d workspace exml lib jeuclid fop jar lib d workspace exml lib jai tiff jar lib d workspace exml lib jh jar lib d workspace dita ot lib jsearch jar lib d workspace exml lib lucene analyzers common jar lib d workspace exml lib lucene analyzers kuromoji jar lib d workspace exml lib lucene core jar lib d workspace exml lib lucene misc jar lib d workspace exml lib lucene queries jar lib d workspace exml lib lucene queryparser jar lib d workspace exml lib lucene suggest jar lib d workspace dita ot lib d workspace dita ot lib dost patches jar lib d workspace dita ot lib lib d workspace dita ot lib ant apache resolver jar lib d workspace dita ot lib ant launcher jar lib d workspace dita ot lib ant jar lib d workspace dita ot lib commons codec jar lib d workspace dita ot lib commons io jar lib d workspace dita ot lib dost configuration jar lib d workspace dita ot lib dost jar lib d workspace dita ot lib guava jar lib d workspace dita ot lib jar lib d workspace dita ot lib saxon dom jar lib d workspace dita ot lib saxon jar lib d workspace dita ot lib xercesimpl jar lib d workspace dita ot lib xml apis jar lib d workspace dita ot lib xml resolver jar lib d workspace dita ot plugins org dita odt lib odt jar lib d workspace dita ot plugins org dita lib fo jar lib d workspace dita ot plugins org dita axf lib axf jar lib d workspace dita ot plugins org dita xep lib xep jar lib d workspace dita ot plugins org dita wordrtf lib wordrtf jar f d workspace dita ot build xml dtranstype xhtml dbasedir c users radu pisoi desktop temp dita issue flowers doutput dir c users radu pisoi desktop temp dita issue flowers out xhtml ddita temp dir c users radu pisoi desktop temp dita issue flowers temp xhtml dargs input c users radu pisoi desktop temp dita issue flowers flowers ditamap ddita dir d workspace dita ot dbasejvmargline buildfile d workspace dita ot build xml init init check arg created dir c users radu pisoi desktop temp dita issue flowers out xhtml created dir c users radu pisoi desktop temp dita issue flowers temp xhtml log arg basedir c users radu pisoi desktop temp dita issue flowers dita dir d workspace dita ot transtype xhtml tempdir c users radu pisoi desktop temp dita issue flowers temp xhtml outputdir c users radu pisoi desktop temp dita issue flowers out xhtml clean temp true dita ot version xml parser xerces xslt processor saxon collator icu ant properties fri aug eest args grammar cache yes args input c users radu pisoi desktop temp dita issue flowers flowers ditamap args logdir c users radu pisoi desktop temp dita issue flowers out xhtml args xml systemid set yes dita dir d workspace dita ot dita plugin com sophos tocjs dir d workspace dita ot plugins com sophos tocjs dita plugin org dita base dir d workspace dita ot dita plugin org dita docbook dir d workspace dita ot plugins org dita docbook dita plugin org dita eclipsecontent dir d workspace dita ot plugins org dita eclipsecontent dita plugin org dita eclipsehelp dir d workspace dita ot plugins org dita eclipsehelp dita plugin org dita dir d workspace dita ot plugins org dita dita plugin org dita htmlhelp dir d workspace dita ot plugins org dita htmlhelp dita plugin org dita javahelp dir d workspace dita ot plugins org dita javahelp dita plugin org dita odt dir d workspace dita ot plugins org dita odt dita plugin org dita axf dir d workspace dita ot plugins org dita axf dita plugin org dita dir d workspace dita ot plugins org dita dita plugin org dita fop dir d workspace dita ot plugins org dita fop dita plugin org dita xep dir d workspace dita ot plugins org dita xep dita plugin org dita specialization dir d workspace dita ot plugins org dita specialization dita plugin org dita specialization eclipsemap dir d workspace dita ot plugins org dita specialization eclipsemap dita plugin org dita troff dir d workspace dita ot plugins org dita troff dita plugin org dita wordrtf dir d workspace dita ot plugins org dita wordrtf dita plugin org dita xhtml dir d workspace dita ot plugins org dita xhtml dita plugin org oasis open dita dir d workspace dita ot plugins org oasis open dita dita plugin org oasis open dita dir d workspace dita ot plugins org oasis open dita dita temp dir c users radu pisoi desktop temp dita issue flowers temp xhtml topicpull check topicpull transforming into c users radu pisoi desktop temp dita issue flowers temp xhtml loading stylesheet d workspace dita ot xsl preprocess topicpull xsl processing c users radu pisoi desktop temp dita issue flowers temp xhtml topics index dita processing c users radu pisoi desktop temp dita issue flowers temp xhtml concepts springflowers dita recoverable error on line of topicpullimpl xsl java io filenotfoundexception c users radu pisoi desktop temp dita issue flowers temp xhtml topics flowers concepts springflowers dita the system cannot find the path specified file c users radu pisoi desktop temp dita issue flowers topics flowers snowdrop dita the file file c users radu pisoi desktop temp dita issue flowers temp xhtml topics flowers concepts springflowers dita is not available to resolve link information recoverable error on line of topicpullimpl xsl java io filenotfoundexception c users radu pisoi desktop temp dita issue flowers temp xhtml topics flowers concepts springflowers dita the system cannot find the path specified recoverable error on line of topicpullimpl xsl java io filenotfoundexception c users radu pisoi desktop temp dita issue flowers temp xhtml topics flowers concepts springflowers dita the system cannot find the path specified recoverable error on line of topicpullimpl xsl java io filenotfoundexception c users radu pisoi desktop temp dita issue flowers temp xhtml topics flowers concepts springflowers dita the system cannot find the path specified i ve tested with dita ot and the attached dita map
1
9,477
12,475,855,306
IssuesEvent
2020-05-29 12:26:31
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Possibility to create vector tiles directly from QGIS
Feature Request Processing Vector tiles
Ideally a processing algorithm (similar to mbtiles/xyz tiles) to generate vector tiles from QGIS map.
1.0
Possibility to create vector tiles directly from QGIS - Ideally a processing algorithm (similar to mbtiles/xyz tiles) to generate vector tiles from QGIS map.
process
possibility to create vector tiles directly from qgis ideally a processing algorithm similar to mbtiles xyz tiles to generate vector tiles from qgis map
1
15,242
19,179,995,635
IssuesEvent
2021-12-04 07:50:01
varabyte/kobweb
https://api.github.com/repos/varabyte/kobweb
closed
Audit ktor plugins
process
When I created the initial project, I selected a bunch of plugins that seemed relevant, but right now they may not be being used, and some of them are generating compiler warnings.
1.0
Audit ktor plugins - When I created the initial project, I selected a bunch of plugins that seemed relevant, but right now they may not be being used, and some of them are generating compiler warnings.
process
audit ktor plugins when i created the initial project i selected a bunch of plugins that seemed relevant but right now they may not be being used and some of them are generating compiler warnings
1
5,952
8,775,486,062
IssuesEvent
2018-12-18 23:19:48
knative/serving
https://api.github.com/repos/knative/serving
closed
Replicate knative-releases to {us,asia,eu}.gcr.io
area/test-and-release kind/process
<!-- Pro-tip: You can leave this block commented, and it still works! Select the appropriate areas for your issue: /area test-and-release Classify what kind of issue this is: /kind process --> ## Expected Behavior Our releases should be replicated to different regions to improve availability and spacial locality. ## Actual Behavior Our releases are published only to gcr.io.
1.0
Replicate knative-releases to {us,asia,eu}.gcr.io - <!-- Pro-tip: You can leave this block commented, and it still works! Select the appropriate areas for your issue: /area test-and-release Classify what kind of issue this is: /kind process --> ## Expected Behavior Our releases should be replicated to different regions to improve availability and spacial locality. ## Actual Behavior Our releases are published only to gcr.io.
process
replicate knative releases to us asia eu gcr io pro tip you can leave this block commented and it still works select the appropriate areas for your issue area test and release classify what kind of issue this is kind process expected behavior our releases should be replicated to different regions to improve availability and spacial locality actual behavior our releases are published only to gcr io
1
5,598
8,459,928,961
IssuesEvent
2018-10-22 17:19:29
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
Change specific responses to chemicals to part_of response to X
cellular processes
During the [June 2014 meeting on annotation extension relations](http://wiki.geneontology.org/index.php/Annotation_Extension_meeting_2014-06-16), it was agreed to change terms for specific responses to chemicals (see [item 1 from 140624 annotation call minutes](http://wiki.geneontology.org/index.php/Annotation_Conf._Call,_June_24,_2014#Summary_of_proposed_action_items_from_annotation_extension_relation_meeting)). Examples: [chemosensory behavior](http://purl.obolibrary.org/obo/GO_0007635) is_a 'response to chemical' -> [chemosensory behavior](http://purl.obolibrary.org/obo/GO_0007635) part_of 'response to chemical' Motivating example (seems to be fictional term): ‘proteolysis [involved] in cellular response to drug' could have two has_input relationships: has_input: drug has_input: proteolysis target. The has_inputs work in the individual cases but when combined, how do you know which input is which? A. The drug isn’t an input to the proteolysis. The proteolysis is part of the cellular response to drug. So 'proteolysis [involved] in cellular response to drug' should be part of cellular response to drug, rather than is_a. Reported by: dosumis Original Ticket: [geneontology/ontology-requests/10958](https://sourceforge.net/p/geneontology/ontology-requests/10958)
1.0
Change specific responses to chemicals to part_of response to X - During the [June 2014 meeting on annotation extension relations](http://wiki.geneontology.org/index.php/Annotation_Extension_meeting_2014-06-16), it was agreed to change terms for specific responses to chemicals (see [item 1 from 140624 annotation call minutes](http://wiki.geneontology.org/index.php/Annotation_Conf._Call,_June_24,_2014#Summary_of_proposed_action_items_from_annotation_extension_relation_meeting)). Examples: [chemosensory behavior](http://purl.obolibrary.org/obo/GO_0007635) is_a 'response to chemical' -> [chemosensory behavior](http://purl.obolibrary.org/obo/GO_0007635) part_of 'response to chemical' Motivating example (seems to be fictional term): ‘proteolysis [involved] in cellular response to drug' could have two has_input relationships: has_input: drug has_input: proteolysis target. The has_inputs work in the individual cases but when combined, how do you know which input is which? A. The drug isn’t an input to the proteolysis. The proteolysis is part of the cellular response to drug. So 'proteolysis [involved] in cellular response to drug' should be part of cellular response to drug, rather than is_a. Reported by: dosumis Original Ticket: [geneontology/ontology-requests/10958](https://sourceforge.net/p/geneontology/ontology-requests/10958)
process
change specific responses to chemicals to part of response to x during the it was agreed to change terms for specific responses to chemicals see examples is a response to chemical part of response to chemical motivating example seems to be fictional term ‘proteolysis in cellular response to drug could have two has input relationships has input drug has input proteolysis target the has inputs work in the individual cases but when combined how do you know which input is which a the drug isn’t an input to the proteolysis the proteolysis is part of the cellular response to drug so proteolysis in cellular response to drug should be part of cellular response to drug rather than is a reported by dosumis original ticket
1
2,505
5,238,993,129
IssuesEvent
2017-01-31 08:08:21
vimperator/vimperator-labs
https://api.github.com/repos/vimperator/vimperator-labs
closed
Vimperator don't work in Firefox Beta. Not e10s.
compatibility
##### Issue type: <!-- Pick one and delete the rest --> - Version compatibility ##### Version: ``` Vimperator 3.15, Nightly 51.0a2, e10s disabled. ``` ##### Description: Mappings and commandline don't work. In browser console it's showing that `ReferenceError: tabs is not defined` https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Errors/Not_defined
True
Vimperator don't work in Firefox Beta. Not e10s. - ##### Issue type: <!-- Pick one and delete the rest --> - Version compatibility ##### Version: ``` Vimperator 3.15, Nightly 51.0a2, e10s disabled. ``` ##### Description: Mappings and commandline don't work. In browser console it's showing that `ReferenceError: tabs is not defined` https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Errors/Not_defined
non_process
vimperator don t work in firefox beta not issue type version compatibility version vimperator nightly disabled description mappings and commandline don t work in browser console it s showing that referenceerror tabs is not defined
0
20,452
6,041,080,163
IssuesEvent
2017-06-10 20:32:30
HopefulLlama/UnitTestSCAD
https://api.github.com/repos/HopefulLlama/UnitTestSCAD
closed
Fix "no-return-assign" issue in src/tester/Assertions.js
codeclimate technical
Return statement should not contain assignment. https://codeclimate.com/github/HopefulLlama/UnitTestSCAD/src/tester/Assertions.js#issue_593b221e9711610001000034
1.0
Fix "no-return-assign" issue in src/tester/Assertions.js - Return statement should not contain assignment. https://codeclimate.com/github/HopefulLlama/UnitTestSCAD/src/tester/Assertions.js#issue_593b221e9711610001000034
non_process
fix no return assign issue in src tester assertions js return statement should not contain assignment
0
14,204
17,102,793,953
IssuesEvent
2021-07-09 13:39:06
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
Status of Bazel 5.0.0-pre.20210604.6
P1 release team-XProduct type: process
- Expected release date: June 11th Task list: - [x] Pick release baseline: d571c8976f279e4d485498d0875c85ae73aecc42 with cherrypicks c367a7e027f351e8ffefabf919563abbb8c7d0e1 260015df1524543d56b0dbff00958320cd58d166 - [x] Create release candidate: https://releases.bazel.build/5.0.0/rolling/5.0.0-pre.20210604.6rc1/index.html - [x] Check Bazel post-submit: https://buildkite.com/bazel/bazel-bazel/builds/16498 - [x] Check downstream projects: https://buildkite.com/bazel/bazel-at-head-plus-downstream/builds/2065 - [x] Push the release: https://releases.bazel.build/5.0.0/rolling/5.0.0-pre.20210604.6/index.html - [x] Update the [release page](https://github.com/bazelbuild/bazel/releases/)
1.0
Status of Bazel 5.0.0-pre.20210604.6 - - Expected release date: June 11th Task list: - [x] Pick release baseline: d571c8976f279e4d485498d0875c85ae73aecc42 with cherrypicks c367a7e027f351e8ffefabf919563abbb8c7d0e1 260015df1524543d56b0dbff00958320cd58d166 - [x] Create release candidate: https://releases.bazel.build/5.0.0/rolling/5.0.0-pre.20210604.6rc1/index.html - [x] Check Bazel post-submit: https://buildkite.com/bazel/bazel-bazel/builds/16498 - [x] Check downstream projects: https://buildkite.com/bazel/bazel-at-head-plus-downstream/builds/2065 - [x] Push the release: https://releases.bazel.build/5.0.0/rolling/5.0.0-pre.20210604.6/index.html - [x] Update the [release page](https://github.com/bazelbuild/bazel/releases/)
process
status of bazel pre expected release date june task list pick release baseline with cherrypicks create release candidate check bazel post submit check downstream projects push the release update the
1
406
2,848,763,691
IssuesEvent
2015-05-30 04:02:19
dita-ot/dita-ot
https://api.github.com/repos/dita-ot/dita-ot
closed
Error reported during flagging [DOT 2.0]
bug P2 preprocess
At some point when publishing DITA content to XHTML using DITA OT 2.0 and the latest Saxon 9.6 I get errors like: [preprocess_flag] Processing C:\Users\radu_coravu\Documents\userguide-private\DITA\temp\webhelp\topics\author-dita-templates.dita [preprocess_flag] Error on line 563 of flagImpl.xsl: [preprocess_flag] XTTE0590: An empty sequence is not allowed as the value of parameter $flag-att-val [preprocess_flag] at xsl:call-template name="gen-prop" (file:/D:/projects/eXml/frameworks/dita/DITA-OT2.x/xsl/preprocess/flagImpl.xsl#541) [preprocess_flag] at xsl:call-template name="ext-gen-prop" (file:/D:/projects/eXml/frameworks/dita/DITA-OT2.x/xsl/preprocess/flagImpl.xsl#415) [preprocess_flag] at xsl:call-template name="ext-getrules" (file:/D:/projects/eXml/frameworks/dita/DITA-OT2.x/xsl/preprocess/flagImpl.xsl#375) [preprocess_flag] at xsl:call-template name="getrules" (file:/D:/projects/eXml/frameworks/dita/DITA-OT2.x/xsl/preprocess/flagImpl.xsl#85) [preprocess_flag] at xsl:apply-templates (file:/D:/projects/eXml/frameworks/dita/DITA-OT2.x/xsl/preprocess/flagImpl.xsl#80) [preprocess_flag] processing /topic [preprocess_flag] Failed to transform document: An empty sequence is not allowed as the value of parameter $flag-att-val At some point in that stylesheet there is a parameter defined as being of type xs:string. <xsl:param name="flag-att-val" as="xs:string"/> But a value for it is passed from this parameter: <xsl:param name="props" as="xs:string?"/> Once you add an occurrence indicator after a type, it becomes a sequence (in our case empty sequence or sequence of a single string): http://www.w3.org/TR/xpath20/#prod-xpath-OccurrenceIndicator so it seems that the latest Saxon 9.6 reports an error and no longer automatically converts the sequence to a plain string when it is passed as a parameter.
1.0
Error reported during flagging [DOT 2.0] - At some point when publishing DITA content to XHTML using DITA OT 2.0 and the latest Saxon 9.6 I get errors like: [preprocess_flag] Processing C:\Users\radu_coravu\Documents\userguide-private\DITA\temp\webhelp\topics\author-dita-templates.dita [preprocess_flag] Error on line 563 of flagImpl.xsl: [preprocess_flag] XTTE0590: An empty sequence is not allowed as the value of parameter $flag-att-val [preprocess_flag] at xsl:call-template name="gen-prop" (file:/D:/projects/eXml/frameworks/dita/DITA-OT2.x/xsl/preprocess/flagImpl.xsl#541) [preprocess_flag] at xsl:call-template name="ext-gen-prop" (file:/D:/projects/eXml/frameworks/dita/DITA-OT2.x/xsl/preprocess/flagImpl.xsl#415) [preprocess_flag] at xsl:call-template name="ext-getrules" (file:/D:/projects/eXml/frameworks/dita/DITA-OT2.x/xsl/preprocess/flagImpl.xsl#375) [preprocess_flag] at xsl:call-template name="getrules" (file:/D:/projects/eXml/frameworks/dita/DITA-OT2.x/xsl/preprocess/flagImpl.xsl#85) [preprocess_flag] at xsl:apply-templates (file:/D:/projects/eXml/frameworks/dita/DITA-OT2.x/xsl/preprocess/flagImpl.xsl#80) [preprocess_flag] processing /topic [preprocess_flag] Failed to transform document: An empty sequence is not allowed as the value of parameter $flag-att-val At some point in that stylesheet there is a parameter defined as being of type xs:string. <xsl:param name="flag-att-val" as="xs:string"/> But a value for it is passed from this parameter: <xsl:param name="props" as="xs:string?"/> Once you add an occurrence indicator after a type, it becomes a sequence (in our case empty sequence or sequence of a single string): http://www.w3.org/TR/xpath20/#prod-xpath-OccurrenceIndicator so it seems that the latest Saxon 9.6 reports an error and no longer automatically converts the sequence to a plain string when it is passed as a parameter.
process
error reported during flagging at some point when publishing dita content to xhtml using dita ot and the latest saxon i get errors like processing c users radu coravu documents userguide private dita temp webhelp topics author dita templates dita error on line of flagimpl xsl an empty sequence is not allowed as the value of parameter flag att val at xsl call template name gen prop file d projects exml frameworks dita dita x xsl preprocess flagimpl xsl at xsl call template name ext gen prop file d projects exml frameworks dita dita x xsl preprocess flagimpl xsl at xsl call template name ext getrules file d projects exml frameworks dita dita x xsl preprocess flagimpl xsl at xsl call template name getrules file d projects exml frameworks dita dita x xsl preprocess flagimpl xsl at xsl apply templates file d projects exml frameworks dita dita x xsl preprocess flagimpl xsl processing topic failed to transform document an empty sequence is not allowed as the value of parameter flag att val at some point in that stylesheet there is a parameter defined as being of type xs string but a value for it is passed from this parameter once you add an occurrence indicator after a type it becomes a sequence in our case empty sequence or sequence of a single string so it seems that the latest saxon reports an error and no longer automatically converts the sequence to a plain string when it is passed as a parameter
1
8,000
11,188,530,197
IssuesEvent
2020-01-02 05:37:52
PHPSocialNetwork/phpfastcache
https://api.github.com/repos/PHPSocialNetwork/phpfastcache
closed
Api Method "deleteItemsByTagsAll()" removes unrelated items
8.0 :( Not fixable >_< Working & Scheduled [-_-] In Process
When You add multiple items that have their own tag and also share a tag it seems that when you delete an item using `deleteItemsByTagsAll(['sharedTag', 'ownTag'])` that all items that share `'sharedTag'` are also removed. Following the docs it should only remove items that have multiple tags or all you set for this function, not one or more like deleteItemsByTags() does. This happens for me on Couchbase buckets. **To Reproduce** Steps to reproduce the behavior: 1. Save multiple items with a shared tag and their own tag (like ID) 2. Remove one of the items by it's IDtag and the sharedTag using `deleteItemsByTagsAll(['sharedTag', 'ownTag'])` So 10 items should be in total 21 => 10 items, 10 unique tags items, 1 shared tag item **Expected behavior** One item removed and one (unique) tag, so 2 items less in the bucket and the rest should stay **Current behavior** All items that share the shared tags are removed, so also are their individual tags.
1.0
Api Method "deleteItemsByTagsAll()" removes unrelated items - When You add multiple items that have their own tag and also share a tag it seems that when you delete an item using `deleteItemsByTagsAll(['sharedTag', 'ownTag'])` that all items that share `'sharedTag'` are also removed. Following the docs it should only remove items that have multiple tags or all you set for this function, not one or more like deleteItemsByTags() does. This happens for me on Couchbase buckets. **To Reproduce** Steps to reproduce the behavior: 1. Save multiple items with a shared tag and their own tag (like ID) 2. Remove one of the items by it's IDtag and the sharedTag using `deleteItemsByTagsAll(['sharedTag', 'ownTag'])` So 10 items should be in total 21 => 10 items, 10 unique tags items, 1 shared tag item **Expected behavior** One item removed and one (unique) tag, so 2 items less in the bucket and the rest should stay **Current behavior** All items that share the shared tags are removed, so also are their individual tags.
process
api method deleteitemsbytagsall removes unrelated items when you add multiple items that have their own tag and also share a tag it seems that when you delete an item using deleteitemsbytagsall that all items that share sharedtag are also removed following the docs it should only remove items that have multiple tags or all you set for this function not one or more like deleteitemsbytags does this happens for me on couchbase buckets to reproduce steps to reproduce the behavior save multiple items with a shared tag and their own tag like id remove one of the items by it s idtag and the sharedtag using deleteitemsbytagsall so items should be in total items unique tags items shared tag item expected behavior one item removed and one unique tag so items less in the bucket and the rest should stay current behavior all items that share the shared tags are removed so also are their individual tags
1
20,074
26,568,442,866
IssuesEvent
2023-01-20 23:09:08
srophe/caesarea-data
https://api.github.com/repos/srophe/caesarea-data
closed
Test out why pipe character conversion to <lb/> isn't working consistently
post-processor
Cf. #150. It seems like it only worked part of the time. Need to dig into this a bit more.
1.0
Test out why pipe character conversion to <lb/> isn't working consistently - Cf. #150. It seems like it only worked part of the time. Need to dig into this a bit more.
process
test out why pipe character conversion to isn t working consistently cf it seems like it only worked part of the time need to dig into this a bit more
1
11,632
14,490,871,169
IssuesEvent
2020-12-11 03:17:05
cypress-io/cypress
https://api.github.com/repos/cypress-io/cypress
reopened
"Found .only in folder" error when trying to commit
process: contributing type: chore
### Current behavior: I tried to commit into my own branch in my fork of cypress. This fails with the error: > Found .only in folder(s) 👎 > 544: * Mocks global clock but only overrides specific functions. > 1847: * Creates a spy / stub but only for calls with given arguments. > 2108: * Whether Cypress will upload the video to the Dashboard even if all tests are passing. This applies only when recording your runs to the Dashboard. Turn this off if you’d like the video uploaded only when there are failing tests. > 2721: * Causes all `.keys` assertions that follow in the chain to only require that the target have at least one of the given keys. This is the opposite of `.all`, which requires that the target have all of the given keys. > 3221: * Causes all `.keys` assertions that follow in the chain to only require that the target not have at least one of the given keys. This is the opposite of `.all`, which requires that the target have all of the given keys. > 4352: /** Override *name* for display purposes only */ But I have done nothing related to this I think. ### Desired behavior: I should be able to commit on Windows without using "--no-verify" ### Steps to reproduce: (app code and test code) Fork cypress-io/cypress repo, make a branch and try to commit the changes made in https://github.com/gabbersepp/cypress/commit/de0a7c0d4936886b1b7576b60bb0b533c9207cb2 ### Versions Current develop branch, Win10
1.0
"Found .only in folder" error when trying to commit - ### Current behavior: I tried to commit into my own branch in my fork of cypress. This fails with the error: > Found .only in folder(s) 👎 > 544: * Mocks global clock but only overrides specific functions. > 1847: * Creates a spy / stub but only for calls with given arguments. > 2108: * Whether Cypress will upload the video to the Dashboard even if all tests are passing. This applies only when recording your runs to the Dashboard. Turn this off if you’d like the video uploaded only when there are failing tests. > 2721: * Causes all `.keys` assertions that follow in the chain to only require that the target have at least one of the given keys. This is the opposite of `.all`, which requires that the target have all of the given keys. > 3221: * Causes all `.keys` assertions that follow in the chain to only require that the target not have at least one of the given keys. This is the opposite of `.all`, which requires that the target have all of the given keys. > 4352: /** Override *name* for display purposes only */ But I have done nothing related to this I think. ### Desired behavior: I should be able to commit on Windows without using "--no-verify" ### Steps to reproduce: (app code and test code) Fork cypress-io/cypress repo, make a branch and try to commit the changes made in https://github.com/gabbersepp/cypress/commit/de0a7c0d4936886b1b7576b60bb0b533c9207cb2 ### Versions Current develop branch, Win10
process
found only in folder error when trying to commit current behavior i tried to commit into my own branch in my fork of cypress this fails with the error found only in folder s 👎 mocks global clock but only overrides specific functions creates a spy stub but only for calls with given arguments whether cypress will upload the video to the dashboard even if all tests are passing this applies only when recording your runs to the dashboard turn this off if you’d like the video uploaded only when there are failing tests causes all keys assertions that follow in the chain to only require that the target have at least one of the given keys this is the opposite of all which requires that the target have all of the given keys causes all keys assertions that follow in the chain to only require that the target not have at least one of the given keys this is the opposite of all which requires that the target have all of the given keys override name for display purposes only but i have done nothing related to this i think desired behavior i should be able to commit on windows without using no verify steps to reproduce app code and test code fork cypress io cypress repo make a branch and try to commit the changes made in versions current develop branch
1
9,499
12,488,728,271
IssuesEvent
2020-05-31 15:33:03
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
Sorting by two columns with the same name fails with ambiguous column error
Priority:P2 Querying/Processor Type:Bug
For example if you create a query that has two `count` columns in the results and then click the column header to sort the resulting query will fail because the sort is ambiguous. This can be confusing because it's not obvious how to remove sorts in the query builder (you need to use the `...` menu even though you didn't need to use it to add the sort). There's a couple things we could do to fix this: 1. disallow sorts for ambiguous columns since we know they'll break 2. try to prevent ambiguous columns from being created in the first place. e.g. by naming our columns `count` and `count_2` in the query instead of having `clojure.java.jdbc` do it for us in the results. 3. let people manually specify column names, perhaps just reusing the existing UI for choosing which columns to display/etc.
1.0
Sorting by two columns with the same name fails with ambiguous column error - For example if you create a query that has two `count` columns in the results and then click the column header to sort the resulting query will fail because the sort is ambiguous. This can be confusing because it's not obvious how to remove sorts in the query builder (you need to use the `...` menu even though you didn't need to use it to add the sort). There's a couple things we could do to fix this: 1. disallow sorts for ambiguous columns since we know they'll break 2. try to prevent ambiguous columns from being created in the first place. e.g. by naming our columns `count` and `count_2` in the query instead of having `clojure.java.jdbc` do it for us in the results. 3. let people manually specify column names, perhaps just reusing the existing UI for choosing which columns to display/etc.
process
sorting by two columns with the same name fails with ambiguous column error for example if you create a query that has two count columns in the results and then click the column header to sort the resulting query will fail because the sort is ambiguous this can be confusing because it s not obvious how to remove sorts in the query builder you need to use the menu even though you didn t need to use it to add the sort there s a couple things we could do to fix this disallow sorts for ambiguous columns since we know they ll break try to prevent ambiguous columns from being created in the first place e g by naming our columns count and count in the query instead of having clojure java jdbc do it for us in the results let people manually specify column names perhaps just reusing the existing ui for choosing which columns to display etc
1
127,515
27,063,260,597
IssuesEvent
2023-02-13 21:36:53
eclipse-theia/theia
https://api.github.com/repos/eclipse-theia/theia
closed
Infer plugin activation events from contributions
plug-in system vscode
### Feature Description: Starting with the newest version of VSCode, the plugin host infers activation events in part from the `package.json` contributions. This is highlighted by a warning in the activation events: ![image](https://user-images.githubusercontent.com/4377073/217531526-8f4d19f0-87a6-48e1-9b7d-79a8a83a4033.png) We should implement a similar mechanism soon, as otherwise newer extensions (with old vscode version requirements) might not be activated as expected.
1.0
Infer plugin activation events from contributions - ### Feature Description: Starting with the newest version of VSCode, the plugin host infers activation events in part from the `package.json` contributions. This is highlighted by a warning in the activation events: ![image](https://user-images.githubusercontent.com/4377073/217531526-8f4d19f0-87a6-48e1-9b7d-79a8a83a4033.png) We should implement a similar mechanism soon, as otherwise newer extensions (with old vscode version requirements) might not be activated as expected.
non_process
infer plugin activation events from contributions feature description starting with the newest version of vscode the plugin host infers activation events in part from the package json contributions this is highlighted by a warning in the activation events we should implement a similar mechanism soon as otherwise newer extensions with old vscode version requirements might not be activated as expected
0
971
2,522,283,567
IssuesEvent
2015-01-19 20:50:42
vanstyn/RapidApp
https://api.github.com/repos/vanstyn/RapidApp
opened
Unit tests for SimpleCAS/CMS uploads
Test Coverage
Using ``RapidApp::Test``, unit tests need to be written to confirm operation of file/image uploads: * Uploading a normal file and getting back link * Access the uploaded file via that link * Uploading an image and getting back link+meta data info (height/width) * Uploading an image that is larger than max size and getting back warning and resized details
1.0
Unit tests for SimpleCAS/CMS uploads - Using ``RapidApp::Test``, unit tests need to be written to confirm operation of file/image uploads: * Uploading a normal file and getting back link * Access the uploaded file via that link * Uploading an image and getting back link+meta data info (height/width) * Uploading an image that is larger than max size and getting back warning and resized details
non_process
unit tests for simplecas cms uploads using rapidapp test unit tests need to be written to confirm operation of file image uploads uploading a normal file and getting back link access the uploaded file via that link uploading an image and getting back link meta data info height width uploading an image that is larger than max size and getting back warning and resized details
0
105,323
11,441,158,043
IssuesEvent
2020-02-05 11:04:54
IlonaMenkui/super-enigma-service
https://api.github.com/repos/IlonaMenkui/super-enigma-service
closed
Write Swagger spec
documentation enhancement
Write Swagger spec about available REST endpoints. Place this doc under `/spec/http/api-v1.yml` folder. Resources: 1. Doc - https://swagger.io/docs/specification/2-0/basic-structure/ 2. Examples - https://github.com/OAI/OpenAPI-Specification/tree/master/examples/v2.0/yaml
1.0
Write Swagger spec - Write Swagger spec about available REST endpoints. Place this doc under `/spec/http/api-v1.yml` folder. Resources: 1. Doc - https://swagger.io/docs/specification/2-0/basic-structure/ 2. Examples - https://github.com/OAI/OpenAPI-Specification/tree/master/examples/v2.0/yaml
non_process
write swagger spec write swagger spec about available rest endpoints place this doc under spec http api yml folder resources doc examples
0
19,626
25,981,257,693
IssuesEvent
2022-12-19 19:04:13
googleapis/google-cloud-go
https://api.github.com/repos/googleapis/google-cloud-go
opened
logging: use custom environment test container
api: logging type: process priority: p3
The logging environment tests are currently running using the same container base as the other go tests. Sometimes, something in the base image shifts, and causes test failures. We should consider creating a new container base image specifically for environment tests. Doing this, we can bake more dependencies into the base image: https://github.com/googleapis/google-cloud-go/pull/7159#discussion_r1052536825
1.0
logging: use custom environment test container - The logging environment tests are currently running using the same container base as the other go tests. Sometimes, something in the base image shifts, and causes test failures. We should consider creating a new container base image specifically for environment tests. Doing this, we can bake more dependencies into the base image: https://github.com/googleapis/google-cloud-go/pull/7159#discussion_r1052536825
process
logging use custom environment test container the logging environment tests are currently running using the same container base as the other go tests sometimes something in the base image shifts and causes test failures we should consider creating a new container base image specifically for environment tests doing this we can bake more dependencies into the base image
1
19,748
26,109,477,253
IssuesEvent
2022-12-27 17:30:57
SigNoz/signoz-otel-collector
https://api.github.com/repos/SigNoz/signoz-otel-collector
closed
Check impact of keeping cache size to 1M or 100M in signozspanmetricsprocessor
signozspanmetricsprocessor
<a href="https://github.com/ankitnayan"><img src="https://avatars.githubusercontent.com/u/12460410?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [ankitnayan](https://github.com/ankitnayan)** _Monday Jul 11, 2022 at 12:09 GMT_ _Originally opened as https://github.com/SigNoz/opentelemetry-collector-contrib/issues/1851_ ---- It is not possible to determine the cache size beforehand as users keep on adding services to signoz. We should have an idea of the impact of keeping a high number for `dimensions_cache_size` like `1M` or `100M`. Or should we change the size dynamically based on some metric/number?
1.0
Check impact of keeping cache size to 1M or 100M in signozspanmetricsprocessor - <a href="https://github.com/ankitnayan"><img src="https://avatars.githubusercontent.com/u/12460410?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [ankitnayan](https://github.com/ankitnayan)** _Monday Jul 11, 2022 at 12:09 GMT_ _Originally opened as https://github.com/SigNoz/opentelemetry-collector-contrib/issues/1851_ ---- It is not possible to determine the cache size beforehand as users keep on adding services to signoz. We should have an idea of the impact of keeping a high number for `dimensions_cache_size` like `1M` or `100M`. Or should we change the size dynamically based on some metric/number?
process
check impact of keeping cache size to or in signozspanmetricsprocessor issue by monday jul at gmt originally opened as it is not possible to determine the cache size beforehand as users keep on adding services to signoz we should have an idea of the impact of keeping a high number for dimensions cache size like or or should we change the size dynamically based on some metric number
1