Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
71,904
| 18,923,046,729
|
IssuesEvent
|
2021-11-17 05:43:44
|
ARM-software/armnn
|
https://api.github.com/repos/ARM-software/armnn
|
closed
|
Compile error for TFlite Delegate
|
Documentation issue Build issue
|
aim at: **Build the TfLite Delegate (Stand-Alone)**
ref from:https://github.com/ARM-software/armnn/blob/branches%2Farmnn_21_02/delegate/BuildGuideNative.md
env:
**1) ubuntu18.04(x86)
2) tensorflow-2.3.1 -build success
3) flatbuffers-1.12.0-build arm64 success
4) compute library_21.02 -build success**
`scons arch=arm64-v8a neon=1 opencl=1 embed_kernels=1 extra_cxx_flags="-fPIC" benchmark_tests=0 validation_tests=0 -j8 internal_only=0`
**5)build armnn 21.02 success**
```
CXX=arrch64-linux-gnu-g++ CC=aarch64-linux-gnu-gcc cmake .. -DARMCOMPUTE_ROOT=$BASEDIR/ComputeLibrary -DARMCOMPUTENEON=1 -DBUILD_UNIT_TESTS=0
make
```
**6) build tflte delegate**
```
CXX=arrch64-linux-gnu-g++ CC=aarch64-linux-gnu-gcc cmake .. -DTENSORFLOW_LIB_DIR=$BASEDIR/tensorflow/bazel-bin -DTENSORFLOW_ROOT=$BASEDIR/tensorflow -DTFLITE_LIB_ROOT=$BASEDIR/tensorflow/bazel-bin -DFLATBUFFERS_ROOT=$BASEDIR/flatbuffers/install -DArmnn_DIR=$BASEDIR/armnn/build -DARMNN_SOURCE_DIR=$BASEDIR/armnn
make
```
/usr/lib/gcc-cross/aarch64-linux-gnu/7/../../../aarch64-linux-gnu/bin/ld:skipping incompatible /home/***/delegate/tensorflow/bazel-bin/libtensorflow_lite_all.so when searching for -ltensorflow_lite_all
/usr/lib/gcc-cross/aarch64-linux-gnu/7/../../../aarch64-linux-gnu/bin/ld: cannot find -ltensorflow_lite_all
look at dirctory,libtensorflow_lite_all.so is exist:
```
ls ../../../tensorflow-bazel-bin
libtensorflow_lite_all.so
```
please give me some advice to slove it,thanks。
|
1.0
|
Compile error for TFlite Delegate - aim at: **Build the TfLite Delegate (Stand-Alone)**
ref from:https://github.com/ARM-software/armnn/blob/branches%2Farmnn_21_02/delegate/BuildGuideNative.md
env:
**1) ubuntu18.04(x86)
2) tensorflow-2.3.1 -build success
3) flatbuffers-1.12.0-build arm64 success
4) compute library_21.02 -build success**
`scons arch=arm64-v8a neon=1 opencl=1 embed_kernels=1 extra_cxx_flags="-fPIC" benchmark_tests=0 validation_tests=0 -j8 internal_only=0`
**5)build armnn 21.02 success**
```
CXX=arrch64-linux-gnu-g++ CC=aarch64-linux-gnu-gcc cmake .. -DARMCOMPUTE_ROOT=$BASEDIR/ComputeLibrary -DARMCOMPUTENEON=1 -DBUILD_UNIT_TESTS=0
make
```
**6) build tflte delegate**
```
CXX=arrch64-linux-gnu-g++ CC=aarch64-linux-gnu-gcc cmake .. -DTENSORFLOW_LIB_DIR=$BASEDIR/tensorflow/bazel-bin -DTENSORFLOW_ROOT=$BASEDIR/tensorflow -DTFLITE_LIB_ROOT=$BASEDIR/tensorflow/bazel-bin -DFLATBUFFERS_ROOT=$BASEDIR/flatbuffers/install -DArmnn_DIR=$BASEDIR/armnn/build -DARMNN_SOURCE_DIR=$BASEDIR/armnn
make
```
/usr/lib/gcc-cross/aarch64-linux-gnu/7/../../../aarch64-linux-gnu/bin/ld:skipping incompatible /home/***/delegate/tensorflow/bazel-bin/libtensorflow_lite_all.so when searching for -ltensorflow_lite_all
/usr/lib/gcc-cross/aarch64-linux-gnu/7/../../../aarch64-linux-gnu/bin/ld: cannot find -ltensorflow_lite_all
look at dirctory,libtensorflow_lite_all.so is exist:
```
ls ../../../tensorflow-bazel-bin
libtensorflow_lite_all.so
```
please give me some advice to slove it,thanks。
|
non_process
|
compile error for tflite delegate aim at build the tflite delegate stand alone ref from env tensorflow build success flatbuffers build success compute library build success scons arch neon opencl embed kernels extra cxx flags fpic benchmark tests validation tests internal only build armnn success cxx linux gnu g cc linux gnu gcc cmake darmcompute root basedir computelibrary darmcomputeneon dbuild unit tests make build tflte delegate cxx linux gnu g cc linux gnu gcc cmake dtensorflow lib dir basedir tensorflow bazel bin dtensorflow root basedir tensorflow dtflite lib root basedir tensorflow bazel bin dflatbuffers root basedir flatbuffers install darmnn dir basedir armnn build darmnn source dir basedir armnn make usr lib gcc cross linux gnu linux gnu bin ld skipping incompatible home delegate tensorflow bazel bin libtensorflow lite all so when searching for ltensorflow lite all usr lib gcc cross linux gnu linux gnu bin ld cannot find ltensorflow lite all look at dirctory,libtensorflow lite all so is exist: ls tensorflow bazel bin libtensorflow lite all so please give me some advice to slove it,thanks。
| 0
|
8,270
| 11,430,788,728
|
IssuesEvent
|
2020-02-04 10:46:25
|
Graylog2/graylog2-server
|
https://api.github.com/repos/Graylog2/graylog2-server
|
closed
|
Grok pattern Extractors crashes and drop logs
|
bug processing
|
After I upgraded Graylog to 3.2 I found that one extractor containing a grok pattern was constantly crashing, the logs were then forever lost.
## Expected Behavior
When I add an extractor with a grok pattern for syslog and select "Always try to extract", graylog should not drop the logs that failed!
## Current Behavior
I added the following extractor :

And I can see in graylog-server logs :
```
2020-02-04T10:16:09.713+01:00 WARN [ProcessBufferProcessor] Unable to process message <fb393704-472e-11ea-bb88-4256e21dfc8e>: java.lang.ClassCastException: Cannot cast java.lang.String to org.joda.time.DateTime
2020-02-04T10:16:09.713+01:00 WARN [ProcessBufferProcessor] Unable to process message <fb393705-472e-11ea-bb88-4256e21dfc8e>: java.lang.ClassCastException: Cannot cast java.lang.String to org.joda.time.DateTime
2020-02-04T10:16:09.713+01:00 WARN [ProcessBufferProcessor] Unable to process message <fb393706-472e-11ea-bb88-4256e21dfc8e>: java.lang.ClassCastException: Cannot cast java.lang.String to org.joda.time.DateTime
2020-02-04T10:16:09.958+01:00 WARN [ProcessBufferProcessor] Unable to process message <fb5e7245-472e-11ea-bb88-4256e21dfc8e>: java.lang.ClassCastException: Cannot cast java.lang.String to org.joda.time.DateTime
2020-02-04T10:16:09.958+01:00 WARN [ProcessBufferProcessor] Unable to process message <fb5e7244-472e-11ea-bb88-4256e21dfc8e>: java.lang.ClassCastException: Cannot cast java.lang.String to org.joda.time.DateTime
2020-02-04T10:16:09.958+01:00 WARN [ProcessBufferProcessor] Unable to process message <fb5e7243-472e-11ea-bb88-4256e21dfc8e>: java.lang.ClassCastException: Cannot cast java.lang.String to org.joda.time.DateTime
2020-02-04T10:16:10.076+01:00 WARN [ProcessBufferProcessor] Unable to process message <fb709ab4-472e-11ea-bb88-4256e21dfc8e>: java.lang.ClassCastException: Cannot cast java.lang.String to org.joda.time.DateTime
2020-02-04T10:16:10.077+01:00 WARN [ProcessBufferProcessor] Unable to process message <fb70c1c0-472e-11ea-bb88-4256e21dfc8e>: java.lang.ClassCastException: Cannot cast java.lang.String to org.joda.time.DateTime
```
When this happened, I have lost a lot of logs that were not indexed in graylog.
When I delete the extractor everything work smoothly.
## Possible Solution
Catch the error so that the logs can be indexed even tho the extractor failed ?
## Steps to Reproduce (for bugs)
1. Graylog in 3.2
2. Add an extractor with the Grok Pattern from above
3. Add logs that do not match the Grok Pattern to Graylog
## Context
This bug automatically dropped logs without warning.
## Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
* Graylog Version: 3.2
* Elasticsearch Version:
* MongoDB Version:
* Operating System: Linux/Debian
* Browser version: Firefox
|
1.0
|
Grok pattern Extractors crashes and drop logs - After I upgraded Graylog to 3.2 I found that one extractor containing a grok pattern was constantly crashing, the logs were then forever lost.
## Expected Behavior
When I add an extractor with a grok pattern for syslog and select "Always try to extract", graylog should not drop the logs that failed!
## Current Behavior
I added the following extractor :

And I can see in graylog-server logs :
```
2020-02-04T10:16:09.713+01:00 WARN [ProcessBufferProcessor] Unable to process message <fb393704-472e-11ea-bb88-4256e21dfc8e>: java.lang.ClassCastException: Cannot cast java.lang.String to org.joda.time.DateTime
2020-02-04T10:16:09.713+01:00 WARN [ProcessBufferProcessor] Unable to process message <fb393705-472e-11ea-bb88-4256e21dfc8e>: java.lang.ClassCastException: Cannot cast java.lang.String to org.joda.time.DateTime
2020-02-04T10:16:09.713+01:00 WARN [ProcessBufferProcessor] Unable to process message <fb393706-472e-11ea-bb88-4256e21dfc8e>: java.lang.ClassCastException: Cannot cast java.lang.String to org.joda.time.DateTime
2020-02-04T10:16:09.958+01:00 WARN [ProcessBufferProcessor] Unable to process message <fb5e7245-472e-11ea-bb88-4256e21dfc8e>: java.lang.ClassCastException: Cannot cast java.lang.String to org.joda.time.DateTime
2020-02-04T10:16:09.958+01:00 WARN [ProcessBufferProcessor] Unable to process message <fb5e7244-472e-11ea-bb88-4256e21dfc8e>: java.lang.ClassCastException: Cannot cast java.lang.String to org.joda.time.DateTime
2020-02-04T10:16:09.958+01:00 WARN [ProcessBufferProcessor] Unable to process message <fb5e7243-472e-11ea-bb88-4256e21dfc8e>: java.lang.ClassCastException: Cannot cast java.lang.String to org.joda.time.DateTime
2020-02-04T10:16:10.076+01:00 WARN [ProcessBufferProcessor] Unable to process message <fb709ab4-472e-11ea-bb88-4256e21dfc8e>: java.lang.ClassCastException: Cannot cast java.lang.String to org.joda.time.DateTime
2020-02-04T10:16:10.077+01:00 WARN [ProcessBufferProcessor] Unable to process message <fb70c1c0-472e-11ea-bb88-4256e21dfc8e>: java.lang.ClassCastException: Cannot cast java.lang.String to org.joda.time.DateTime
```
When this happened, I have lost a lot of logs that were not indexed in graylog.
When I delete the extractor everything work smoothly.
## Possible Solution
Catch the error so that the logs can be indexed even tho the extractor failed ?
## Steps to Reproduce (for bugs)
1. Graylog in 3.2
2. Add an extractor with the Grok Pattern from above
3. Add logs that do not match the Grok Pattern to Graylog
## Context
This bug automatically dropped logs without warning.
## Your Environment
<!--- Include as many relevant details about the environment you experienced the bug in -->
* Graylog Version: 3.2
* Elasticsearch Version:
* MongoDB Version:
* Operating System: Linux/Debian
* Browser version: Firefox
|
process
|
grok pattern extractors crashes and drop logs after i upgraded graylog to i found that one extractor containing a grok pattern was constantly crashing the logs were then forever lost expected behavior when i add an extractor with a grok pattern for syslog and select always try to extract graylog should not drop the logs that failed current behavior i added the following extractor and i can see in graylog server logs warn unable to process message java lang classcastexception cannot cast java lang string to org joda time datetime warn unable to process message java lang classcastexception cannot cast java lang string to org joda time datetime warn unable to process message java lang classcastexception cannot cast java lang string to org joda time datetime warn unable to process message java lang classcastexception cannot cast java lang string to org joda time datetime warn unable to process message java lang classcastexception cannot cast java lang string to org joda time datetime warn unable to process message java lang classcastexception cannot cast java lang string to org joda time datetime warn unable to process message java lang classcastexception cannot cast java lang string to org joda time datetime warn unable to process message java lang classcastexception cannot cast java lang string to org joda time datetime when this happened i have lost a lot of logs that were not indexed in graylog when i delete the extractor everything work smoothly possible solution catch the error so that the logs can be indexed even tho the extractor failed steps to reproduce for bugs graylog in add an extractor with the grok pattern from above add logs that do not match the grok pattern to graylog context this bug automatically dropped logs without warning your environment graylog version elasticsearch version mongodb version operating system linux debian browser version firefox
| 1
|
8,542
| 11,714,422,489
|
IssuesEvent
|
2020-03-09 12:19:13
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
docs: needs clarification: "terminal raw mode" (signal handling)
|
doc process tty
|
About the `SIGINT` event on `process` it is said in the docs that
> It is not generated when terminal raw mode is enabled.
https://github.com/nodejs/node/blame/3ec4b21b1c438255df6f1652377011080dc28052/doc/api/process.md#L504
In the example program I play around with the SIGINT handler was not firing. I assumed that I was testing with the "terminal raw mode", so I was looking into understanding what that is and how to disable it.
A web search for `nodejs "terminal raw mode"` didn't yield anything useful, though. I also explored `node --help` and didn't see anything obvious.
I think we should clarify in docs what "terminal raw mode" is and then cross-link to that place from the SIGINT doc I linked above.
|
1.0
|
docs: needs clarification: "terminal raw mode" (signal handling) - About the `SIGINT` event on `process` it is said in the docs that
> It is not generated when terminal raw mode is enabled.
https://github.com/nodejs/node/blame/3ec4b21b1c438255df6f1652377011080dc28052/doc/api/process.md#L504
In the example program I play around with the SIGINT handler was not firing. I assumed that I was testing with the "terminal raw mode", so I was looking into understanding what that is and how to disable it.
A web search for `nodejs "terminal raw mode"` didn't yield anything useful, though. I also explored `node --help` and didn't see anything obvious.
I think we should clarify in docs what "terminal raw mode" is and then cross-link to that place from the SIGINT doc I linked above.
|
process
|
docs needs clarification terminal raw mode signal handling about the sigint event on process it is said in the docs that it is not generated when terminal raw mode is enabled in the example program i play around with the sigint handler was not firing i assumed that i was testing with the terminal raw mode so i was looking into understanding what that is and how to disable it a web search for nodejs terminal raw mode didn t yield anything useful though i also explored node help and didn t see anything obvious i think we should clarify in docs what terminal raw mode is and then cross link to that place from the sigint doc i linked above
| 1
|
2,634
| 5,412,258,728
|
IssuesEvent
|
2017-03-01 14:07:31
|
neuropoly/spinalcordtoolbox
|
https://api.github.com/repos/neuropoly/spinalcordtoolbox
|
opened
|
NURBS are unstable when calculated in voxel coordinates
|
bug fix:minor priority: high sct_process_segmentation
|
especially when voxel spacing in one direction is very large.
This instability causes fluctuations of the derivatives, which induces wrong calculations of the angles. This has an strong effect on CSA computation, as demonstrated in the example below.
Data: ` `
Command:
```
sct_process_segmentation -i t2s_segm.nii.gz -p csa
```
Results:
```
# Slice (z),CSA (mm^2),Angle with respect to the I-S direction (degrees)
0,63.5923927147,18.6837707469
1,79.2658753014,8.26795055912
2,88.2874975915,5.12747654869
3,87.9995780207,6.03282331933
4,-8.61862123509,95.2875209491
5,89.2278470889,4.91847828944
6,95.8647986243,9.45846253778
7,84.0700371176,4.83530051485
8,84.7877521872,1.7651865876
9,81.1628827015,0.530803473898
10,82.0784541216,0.515470049064
11,80.383388821,1.2822922865
12,83.1066671944,1.84459163675
```
|
1.0
|
NURBS are unstable when calculated in voxel coordinates - especially when voxel spacing in one direction is very large.
This instability causes fluctuations of the derivatives, which induces wrong calculations of the angles. This has an strong effect on CSA computation, as demonstrated in the example below.
Data: ` `
Command:
```
sct_process_segmentation -i t2s_segm.nii.gz -p csa
```
Results:
```
# Slice (z),CSA (mm^2),Angle with respect to the I-S direction (degrees)
0,63.5923927147,18.6837707469
1,79.2658753014,8.26795055912
2,88.2874975915,5.12747654869
3,87.9995780207,6.03282331933
4,-8.61862123509,95.2875209491
5,89.2278470889,4.91847828944
6,95.8647986243,9.45846253778
7,84.0700371176,4.83530051485
8,84.7877521872,1.7651865876
9,81.1628827015,0.530803473898
10,82.0784541216,0.515470049064
11,80.383388821,1.2822922865
12,83.1066671944,1.84459163675
```
|
process
|
nurbs are unstable when calculated in voxel coordinates especially when voxel spacing in one direction is very large this instability causes fluctuations of the derivatives which induces wrong calculations of the angles this has an strong effect on csa computation as demonstrated in the example below data command sct process segmentation i segm nii gz p csa results slice z csa mm angle with respect to the i s direction degrees
| 1
|
14,811
| 18,143,492,276
|
IssuesEvent
|
2021-09-25 02:39:25
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Crash when clipping a larger layer with a small layer
|
Feedback stale Processing Bug Crash/Data Corruption Upstream
|
### What is the bug or the crash?
Qgis crashes after approx 10s when clipping a larger layer with a small overlay layer. Restarting Qgis or rebooting the computer doesn't help.

Input layer (Feature count: 2.923.688, CRS ESRI:54009 - World_Mollweide):
[GHS_SMOD_Polygonized_raster2vector.zip](https://drive.google.com/file/d/1QJqkm585TIgtG7J8skGN3nLj7E7SLELO/view?usp=sharing)
Overlay layer (Feature count: 1, CRS EPSG:4326 - WGS 84 ):
[smod_polygonized_raster2vector_smallMaskLayer.zip](https://github.com/qgis/QGIS/files/6960914/smod_polygonized_raster2vector_smallMaskLayer.zip)
### Steps to reproduce the issue
1. Import both layers into a new qgis project
2. processing tools -> clip
3. choose layers according to the screenshot
4. a crash occurs after several seconds
### Versions
QGIS version | 3.20.1-Odense | QGIS code revision | 1c3c5cd6
-- | -- | -- | --
Qt version | 5.15.2
Python version | 3.9.5
GDAL/OGR version | 3.3.1
PROJ version | 8.1.0
EPSG Registry database version | v10.027 (2021-06-17)
GEOS version | 3.9.1-CAPI-1.14.2
SQLite version | 3.35.2
PDAL version | 2.3.0
PostgreSQL client version | 13.0
SpatiaLite version | 5.0.1
QWT version | 6.1.3
QScintilla2 version | 2.11.5
OS version | Windows 10 Version 2009
| | |
Active Python plugins | GroupStatslatlontoolsnominatimQuickOSMdb_managerprocessing
### Additional context
_No response_
|
1.0
|
Crash when clipping a larger layer with a small layer - ### What is the bug or the crash?
Qgis crashes after approx 10s when clipping a larger layer with a small overlay layer. Restarting Qgis or rebooting the computer doesn't help.

Input layer (Feature count: 2.923.688, CRS ESRI:54009 - World_Mollweide):
[GHS_SMOD_Polygonized_raster2vector.zip](https://drive.google.com/file/d/1QJqkm585TIgtG7J8skGN3nLj7E7SLELO/view?usp=sharing)
Overlay layer (Feature count: 1, CRS EPSG:4326 - WGS 84 ):
[smod_polygonized_raster2vector_smallMaskLayer.zip](https://github.com/qgis/QGIS/files/6960914/smod_polygonized_raster2vector_smallMaskLayer.zip)
### Steps to reproduce the issue
1. Import both layers into a new qgis project
2. processing tools -> clip
3. choose layers according to the screenshot
4. a crash occurs after several seconds
### Versions
QGIS version | 3.20.1-Odense | QGIS code revision | 1c3c5cd6
-- | -- | -- | --
Qt version | 5.15.2
Python version | 3.9.5
GDAL/OGR version | 3.3.1
PROJ version | 8.1.0
EPSG Registry database version | v10.027 (2021-06-17)
GEOS version | 3.9.1-CAPI-1.14.2
SQLite version | 3.35.2
PDAL version | 2.3.0
PostgreSQL client version | 13.0
SpatiaLite version | 5.0.1
QWT version | 6.1.3
QScintilla2 version | 2.11.5
OS version | Windows 10 Version 2009
| | |
Active Python plugins | GroupStatslatlontoolsnominatimQuickOSMdb_managerprocessing
### Additional context
_No response_
|
process
|
crash when clipping a larger layer with a small layer what is the bug or the crash qgis crashes after approx when clipping a larger layer with a small overlay layer restarting qgis or rebooting the computer doesn t help input layer feature count crs esri world mollweide overlay layer feature count crs epsg wgs steps to reproduce the issue import both layers into a new qgis project processing tools clip choose layers according to the screenshot a crash occurs after several seconds versions qgis version odense qgis code revision qt version python version gdal ogr version proj version epsg registry database version geos version capi sqlite version pdal version postgresql client version spatialite version qwt version version os version windows version active python plugins groupstatslatlontoolsnominatimquickosmdb managerprocessing additional context no response
| 1
|
100,908
| 30,813,173,077
|
IssuesEvent
|
2023-08-01 11:51:38
|
assistant-ai/jess
|
https://api.github.com/repos/assistant-ai/jess
|
reopened
|
Cloudbuild: use Windows container for `go build` with `CGO_ENABLED=1`
|
feature build
|
Maybe it is the best container. )
https://hub.docker.com/layers/library/golang/windowsservercore-ltsc2022/images/sha256-647b841b8cc8b449ebd00e2774b7fcc8753d7053dd83227c11c306d956662f00?context=explore
|
1.0
|
Cloudbuild: use Windows container for `go build` with `CGO_ENABLED=1` - Maybe it is the best container. )
https://hub.docker.com/layers/library/golang/windowsservercore-ltsc2022/images/sha256-647b841b8cc8b449ebd00e2774b7fcc8753d7053dd83227c11c306d956662f00?context=explore
|
non_process
|
cloudbuild use windows container for go build with cgo enabled maybe it is the best container
| 0
|
303,720
| 23,037,141,865
|
IssuesEvent
|
2022-07-22 20:14:23
|
stoplightio/prism
|
https://api.github.com/repos/stoplightio/prism
|
closed
|
Add protocol requirement to prism proxy URLs
|
documentation
|
This topic:
https://meta.stoplight.io/docs/prism/72d69fb629de0-validation-proxy#validation-proxy
Has an incorrect example.
prism proxy reference/backend/openapi.yaml localhost:3000 --errors
Should be:
prism proxy reference/backend/openapi.yaml http://localhost:3000 --errors
|
1.0
|
Add protocol requirement to prism proxy URLs - This topic:
https://meta.stoplight.io/docs/prism/72d69fb629de0-validation-proxy#validation-proxy
Has an incorrect example.
prism proxy reference/backend/openapi.yaml localhost:3000 --errors
Should be:
prism proxy reference/backend/openapi.yaml http://localhost:3000 --errors
|
non_process
|
add protocol requirement to prism proxy urls this topic has an incorrect example prism proxy reference backend openapi yaml localhost errors should be prism proxy reference backend openapi yaml errors
| 0
|
16,511
| 21,519,625,306
|
IssuesEvent
|
2022-04-28 13:10:55
|
symfony/symfony
|
https://api.github.com/repos/symfony/symfony
|
closed
|
[Process] Allow running multiple commands at once
|
Feature Process Stalled
|
**Description**
When running a lot of small shell commands after each other, the overhead of creating the underlying processes becomes noticeable. It would be handy if the process component would support chaining commands.
Working directly on a Linux/Windows shell you would typically do this:
```shell
# Linux
$ foo; bar; baz
$ foo && bar && baz
# Windows
$ foo & bar & baz
$ foo && bar && baz
```
(Yes there are other chaining operators as well like `||`.)
This is currently not possible with the process component.
**API Example**
I like that `Symfony\Component\Process\Process` became somewhat immutable now (the command line is baked in after the constructor ran), so the API could maybe look like this:
```php
$process = new Process(['foo']);
// foo; bar
$process = $process->withChainedCommand(['bar']); // returns a new process instance
// foo && bar
$process = $process->withAndChainedCommand(['bar']); // returns a new process instance
```
What do you think? Is this a worthwhile addition?
|
1.0
|
[Process] Allow running multiple commands at once - **Description**
When running a lot of small shell commands after each other, the overhead of creating the underlying processes becomes noticeable. It would be handy if the process component would support chaining commands.
Working directly on a Linux/Windows shell you would typically do this:
```shell
# Linux
$ foo; bar; baz
$ foo && bar && baz
# Windows
$ foo & bar & baz
$ foo && bar && baz
```
(Yes there are other chaining operators as well like `||`.)
This is currently not possible with the process component.
**API Example**
I like that `Symfony\Component\Process\Process` became somewhat immutable now (the command line is baked in after the constructor ran), so the API could maybe look like this:
```php
$process = new Process(['foo']);
// foo; bar
$process = $process->withChainedCommand(['bar']); // returns a new process instance
// foo && bar
$process = $process->withAndChainedCommand(['bar']); // returns a new process instance
```
What do you think? Is this a worthwhile addition?
|
process
|
allow running multiple commands at once description when running a lot of small shell commands after each other the overhead of creating the underlying processes becomes noticeable it would be handy if the process component would support chaining commands working directly on a linux windows shell you would typically do this shell linux foo bar baz foo bar baz windows foo bar baz foo bar baz yes there are other chaining operators as well like this is currently not possible with the process component api example i like that symfony component process process became somewhat immutable now the command line is baked in after the constructor ran so the api could maybe look like this php process new process foo bar process process withchainedcommand returns a new process instance foo bar process process withandchainedcommand returns a new process instance what do you think is this a worthwhile addition
| 1
|
67,592
| 27,958,769,316
|
IssuesEvent
|
2023-03-24 14:14:59
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
Chat Completion API
|
cognitive-services/svc triaged cxp doc-enhancement Pri1
|
Hello, I would like to know if "Chat Completion API" means "an API for [chat completion](https://platform.openai.com/docs/guides/chat)" (not a proper noun), or if it's the name of an API, which will remain in English, even for the other languages (which means this "Chat Completion API" name should not be translated). Thanks!
Best regards,
François
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: 13994f3e-a96a-8804-4deb-7e0adad7f479
* Version Independent ID: e087d8ce-bf3a-f12c-41dc-3ea924c6f089
* Content: [How to work with the ChatGPT and GPT-4 models (preview) - Azure OpenAI Service](https://learn.microsoft.com/en-us/azure/cognitive-services/openai/how-to/chatgpt?pivots=programming-language-chat-completions#working-with-the-chat-completion-api)
* Content Source: [articles/cognitive-services/openai/how-to/chatgpt.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/cognitive-services/openai/how-to/chatgpt.md)
* Service: **cognitive-services**
* GitHub Login: @mrbullwinkle
* Microsoft Alias: **mbullwin**
|
1.0
|
Chat Completion API -
Hello, I would like to know if "Chat Completion API" means "an API for [chat completion](https://platform.openai.com/docs/guides/chat)" (not a proper noun), or if it's the name of an API, which will remain in English, even for the other languages (which means this "Chat Completion API" name should not be translated). Thanks!
Best regards,
François
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: 13994f3e-a96a-8804-4deb-7e0adad7f479
* Version Independent ID: e087d8ce-bf3a-f12c-41dc-3ea924c6f089
* Content: [How to work with the ChatGPT and GPT-4 models (preview) - Azure OpenAI Service](https://learn.microsoft.com/en-us/azure/cognitive-services/openai/how-to/chatgpt?pivots=programming-language-chat-completions#working-with-the-chat-completion-api)
* Content Source: [articles/cognitive-services/openai/how-to/chatgpt.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/cognitive-services/openai/how-to/chatgpt.md)
* Service: **cognitive-services**
* GitHub Login: @mrbullwinkle
* Microsoft Alias: **mbullwin**
|
non_process
|
chat completion api hello i would like to know if chat completion api means an api for not a proper noun or if it s the name of an api which will remain in english even for the other languages which means this chat completion api name should not be translated thanks best regards françois document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id version independent id content content source service cognitive services github login mrbullwinkle microsoft alias mbullwin
| 0
|
22,198
| 30,755,628,325
|
IssuesEvent
|
2023-07-29 02:40:47
|
open-telemetry/opentelemetry-collector-contrib
|
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
|
closed
|
[processor/resourcedetection] No system attributes set if `host.id` cannot be fetched
|
bug priority:p1 processor/resourcedetection
|
### Component(s)
_No response_
### What happened?
After https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/24239, instead of setting empty `host.id`, we started throwing an error and dropping all other important system detector attributes: `host.name`, `os.type`, `host.arch`.
We need set other system resources even if `host.id` cannot be fetched.
Also, it's happening even if the `host.id` is disabled, which is now the default behavior. https://github.com/open-telemetry/opentelemetry-collector-contrib/issues/24369 will solve this particular issue. But it's not as important as this one.
### Collector version
0.82.0 (binaries are not released yet)
|
1.0
|
[processor/resourcedetection] No system attributes set if `host.id` cannot be fetched - ### Component(s)
_No response_
### What happened?
After https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/24239, instead of setting empty `host.id`, we started throwing an error and dropping all other important system detector attributes: `host.name`, `os.type`, `host.arch`.
We need set other system resources even if `host.id` cannot be fetched.
Also, it's happening even if the `host.id` is disabled, which is now the default behavior. https://github.com/open-telemetry/opentelemetry-collector-contrib/issues/24369 will solve this particular issue. But it's not as important as this one.
### Collector version
0.82.0 (binaries are not released yet)
|
process
|
no system attributes set if host id cannot be fetched component s no response what happened after instead of setting empty host id we started throwing an error and dropping all other important system detector attributes host name os type host arch we need set other system resources even if host id cannot be fetched also it s happening even if the host id is disabled which is now the default behavior will solve this particular issue but it s not as important as this one collector version binaries are not released yet
| 1
|
12,209
| 14,742,829,363
|
IssuesEvent
|
2021-01-07 12:58:01
|
kdjstudios/SABillingGitlab
|
https://api.github.com/repos/kdjstudios/SABillingGitlab
|
closed
|
68 Portland - SAB Latency
|
anc-process anp-2 ant-support
|
In GitLab by @kdjstudios on Jun 7, 2019, 09:22
**Submitted by:** "Grant Crymes" <grant.crymes@answernet.com>
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/8513066 - CLOSED
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/10208596
**Server:** Internal
**Client/Site:** Portland
**Account:** NA
**Issue:**
We are experiencing latency when posting checks in SAB. It’s taking approximately 8-20 seconds to advance through each screen which is really slowing the payment processing down. This latency issue started a month or so ago but has been getting progressively worse.
|
1.0
|
68 Portland - SAB Latency - In GitLab by @kdjstudios on Jun 7, 2019, 09:22
**Submitted by:** "Grant Crymes" <grant.crymes@answernet.com>
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/8513066 - CLOSED
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/10208596
**Server:** Internal
**Client/Site:** Portland
**Account:** NA
**Issue:**
We are experiencing latency when posting checks in SAB. It’s taking approximately 8-20 seconds to advance through each screen which is really slowing the payment processing down. This latency issue started a month or so ago but has been getting progressively worse.
|
process
|
portland sab latency in gitlab by kdjstudios on jun submitted by grant crymes helpdesk closed helpdesk server internal client site portland account na issue we are experiencing latency when posting checks in sab it’s taking approximately seconds to advance through each screen which is really slowing the payment processing down this latency issue started a month or so ago but has been getting progressively worse
| 1
|
22,089
| 7,113,517,689
|
IssuesEvent
|
2018-01-17 20:46:12
|
dotnet/buildtools
|
https://api.github.com/repos/dotnet/buildtools
|
closed
|
PushToBlobFeed manifest creation should use VSTS default variables from env as a fallback
|
area-building-support
|
For Core repos (and maybe more) these values don't match what we need, but adding this fallback would make the implementation in some places have less code. (See https://github.com/dotnet/core-eng/issues/2404#issuecomment-358389341.)
https://docs.microsoft.com/en-us/vsts/build-release/concepts/definitions/build/variables
* `BUILD_REPOSITORY_NAME`
* `BUILD_SOURCEVERSION`
* `BUILD_SOURCEBRANCHNAME`
* In my experience this has chopped off the `release` part of `release/2.0` so it might not work. `BUILD_SOURCEBRANCH` contains more info than needed but might be better.
* (Edit) Confirmed that in a CoreCLR `release/2.0.0` build, this var is `2.0.0`. SOURCEBRANCH is `refs/heads/release/2.0.0`.
* `BUILD_BUILDID`
/cc @tmat @tannergooding
|
1.0
|
PushToBlobFeed manifest creation should use VSTS default variables from env as a fallback - For Core repos (and maybe more) these values don't match what we need, but adding this fallback would make the implementation in some places have less code. (See https://github.com/dotnet/core-eng/issues/2404#issuecomment-358389341.)
https://docs.microsoft.com/en-us/vsts/build-release/concepts/definitions/build/variables
* `BUILD_REPOSITORY_NAME`
* `BUILD_SOURCEVERSION`
* `BUILD_SOURCEBRANCHNAME`
* In my experience this has chopped off the `release` part of `release/2.0` so it might not work. `BUILD_SOURCEBRANCH` contains more info than needed but might be better.
* (Edit) Confirmed that in a CoreCLR `release/2.0.0` build, this var is `2.0.0`. SOURCEBRANCH is `refs/heads/release/2.0.0`.
* `BUILD_BUILDID`
/cc @tmat @tannergooding
|
non_process
|
pushtoblobfeed manifest creation should use vsts default variables from env as a fallback for core repos and maybe more these values don t match what we need but adding this fallback would make the implementation in some places have less code see build repository name build sourceversion build sourcebranchname in my experience this has chopped off the release part of release so it might not work build sourcebranch contains more info than needed but might be better edit confirmed that in a coreclr release build this var is sourcebranch is refs heads release build buildid cc tmat tannergooding
| 0
|
170,768
| 14,269,525,516
|
IssuesEvent
|
2020-11-21 01:51:17
|
AzureAD/microsoft-authentication-library-for-js
|
https://api.github.com/repos/AzureAD/microsoft-authentication-library-for-js
|
opened
|
Modified copy of msal-angular code inside samples/msal-angular-v2-samples/angular10-browser-sample?
|
documentation question
|
## Library
- [x] `@azure/msal-browser@2.x.x`
- [x] `@azure/msal-angular@2.x.x`
## Documentation location
- [x] Documentation does not exist
## Description
The Angular 10 example does not use the code from `@azure/msal-angular@2.x.x`. Instead it has copied and modified the code, see https://github.com/AzureAD/microsoft-authentication-library-for-js/tree/dev/samples/msal-angular-v2-samples/angular10-browser-sample/src/app/msal.
Why is that? Does it mean we can't use the official `@azure/msal-angular@2.x.x` library, but instead have to copy the code from the example? What happens with bug fixes? Will they be committed to both, @azure/msal-angular@2.x.x` and the example project?
|
1.0
|
Modified copy of msal-angular code inside samples/msal-angular-v2-samples/angular10-browser-sample? - ## Library
- [x] `@azure/msal-browser@2.x.x`
- [x] `@azure/msal-angular@2.x.x`
## Documentation location
- [x] Documentation does not exist
## Description
The Angular 10 example does not use the code from `@azure/msal-angular@2.x.x`. Instead it has copied and modified the code, see https://github.com/AzureAD/microsoft-authentication-library-for-js/tree/dev/samples/msal-angular-v2-samples/angular10-browser-sample/src/app/msal.
Why is that? Does it mean we can't use the official `@azure/msal-angular@2.x.x` library, but instead have to copy the code from the example? What happens with bug fixes? Will they be committed to both, @azure/msal-angular@2.x.x` and the example project?
|
non_process
|
modified copy of msal angular code inside samples msal angular samples browser sample library azure msal browser x x azure msal angular x x documentation location documentation does not exist description the angular example does not use the code from azure msal angular x x instead it has copied and modified the code see why is that does it mean we can t use the official azure msal angular x x library but instead have to copy the code from the example what happens with bug fixes will they be committed to both azure msal angular x x and the example project
| 0
|
22,552
| 31,761,984,305
|
IssuesEvent
|
2023-09-12 06:09:47
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
[Mirror] rules_graalvm@v0.10.2
|
P2 type: process team-OSS mirror request
|
### Please list the URLs of the archives you'd like to mirror:
https://github.com/sgammon/rules_graalvm/releases/download/v0.10.2/rules_graalvm-0.10.2.zip
https://github.com/graalvm/graalvm-ce-builds/releases/download/graal-23.0.1/native-image-installable-svm-java20-windows-amd64-23.0.1.jar
https://github.com/graalvm/graalvm-ce-builds/releases/download/jdk-20.0.2/graalvm-community-jdk-20.0.2_windows-x64_bin.zip
https://github.com/graalvm/graalvm-ce-builds/releases/download/graal-23.0.1/native-image-installable-svm-java20-linux-amd64-23.0.1.jar
https://github.com/graalvm/graalvm-ce-builds/releases/download/jdk-20.0.2/graalvm-community-jdk-20.0.2_linux-x64_bin.tar.gz
https://github.com/graalvm/graalvm-ce-builds/releases/download/graal-23.0.1/native-image-installable-svm-java20-darwin-aarch64-23.0.1.jar
https://github.com/graalvm/graalvm-ce-builds/releases/download/jdk-20.0.2/graalvm-community-jdk-20.0.2_macos-aarch64_bin.tar.gz
https://github.com/graalvm/graalvm-ce-builds/releases/download/graal-23.0.1/native-image-installable-svm-java20-darwin-amd64-23.0.1.jar
https://github.com/graalvm/graalvm-ce-builds/releases/download/jdk-20.0.2/graalvm-community-jdk-20.0.2_macos-x64_bin.tar.gz
|
1.0
|
[Mirror] rules_graalvm@v0.10.2 - ### Please list the URLs of the archives you'd like to mirror:
https://github.com/sgammon/rules_graalvm/releases/download/v0.10.2/rules_graalvm-0.10.2.zip
https://github.com/graalvm/graalvm-ce-builds/releases/download/graal-23.0.1/native-image-installable-svm-java20-windows-amd64-23.0.1.jar
https://github.com/graalvm/graalvm-ce-builds/releases/download/jdk-20.0.2/graalvm-community-jdk-20.0.2_windows-x64_bin.zip
https://github.com/graalvm/graalvm-ce-builds/releases/download/graal-23.0.1/native-image-installable-svm-java20-linux-amd64-23.0.1.jar
https://github.com/graalvm/graalvm-ce-builds/releases/download/jdk-20.0.2/graalvm-community-jdk-20.0.2_linux-x64_bin.tar.gz
https://github.com/graalvm/graalvm-ce-builds/releases/download/graal-23.0.1/native-image-installable-svm-java20-darwin-aarch64-23.0.1.jar
https://github.com/graalvm/graalvm-ce-builds/releases/download/jdk-20.0.2/graalvm-community-jdk-20.0.2_macos-aarch64_bin.tar.gz
https://github.com/graalvm/graalvm-ce-builds/releases/download/graal-23.0.1/native-image-installable-svm-java20-darwin-amd64-23.0.1.jar
https://github.com/graalvm/graalvm-ce-builds/releases/download/jdk-20.0.2/graalvm-community-jdk-20.0.2_macos-x64_bin.tar.gz
|
process
|
rules graalvm please list the urls of the archives you d like to mirror
| 1
|
2,878
| 5,833,394,325
|
IssuesEvent
|
2017-05-09 01:24:04
|
codefordenver/org
|
https://api.github.com/repos/codefordenver/org
|
closed
|
Create Mission Statement and upload to all appropriate places
|
Process
|
Create mission statement to aid fundraising marketing campaign and general understanding of organization.
Get sign-off from heads.
|
1.0
|
Create Mission Statement and upload to all appropriate places - Create mission statement to aid fundraising marketing campaign and general understanding of organization.
Get sign-off from heads.
|
process
|
create mission statement and upload to all appropriate places create mission statement to aid fundraising marketing campaign and general understanding of organization get sign off from heads
| 1
|
133,771
| 18,353,242,454
|
IssuesEvent
|
2021-10-08 14:51:52
|
symfony/symfony-docs
|
https://api.github.com/repos/symfony/symfony-docs
|
closed
|
[Security] Automatically register custom authenticator as entry_point (…
|
Security
|
| Q | A
| ------------ | ---
| Feature PR | symfony/symfony#39153
| PR author(s) | @wouterj
| Merged in | 5.2
|
True
|
[Security] Automatically register custom authenticator as entry_point (… - | Q | A
| ------------ | ---
| Feature PR | symfony/symfony#39153
| PR author(s) | @wouterj
| Merged in | 5.2
|
non_process
|
automatically register custom authenticator as entry point … q a feature pr symfony symfony pr author s wouterj merged in
| 0
|
216,706
| 7,311,092,383
|
IssuesEvent
|
2018-02-28 16:44:53
|
EthereumCommonwealth/etherwallet
|
https://api.github.com/repos/EthereumCommonwealth/etherwallet
|
closed
|
Add more networks.
|
low_priority
|
### Description
Just add custom nodes from https://github.com/kvhnuke/etherwallet/blob/mercury/app/scripts/nodes.js
- Tomo Coin network.
- Ella network.
- POA network.
NOTE: It may be necessary to configure corresponding colorings for each of the networks.
|
1.0
|
Add more networks. - ### Description
Just add custom nodes from https://github.com/kvhnuke/etherwallet/blob/mercury/app/scripts/nodes.js
- Tomo Coin network.
- Ella network.
- POA network.
NOTE: It may be necessary to configure corresponding colorings for each of the networks.
|
non_process
|
add more networks description just add custom nodes from tomo coin network ella network poa network note it may be necessary to configure corresponding colorings for each of the networks
| 0
|
178,030
| 29,486,349,611
|
IssuesEvent
|
2023-06-02 10:01:28
|
OfficeDev/TeamsFx
|
https://api.github.com/repos/OfficeDev/TeamsFx
|
closed
|
TeamsFx spfx tab project Env's replicating each other
|
needs more info *as-designed TA:E2E no recent activity
|
**Describe the bug**
Hi, I have created a spfx tab project from teamfx cli. i have created two env's dev & test which i am managing from teams toolkit exstension. when i deploy new update in dev env and approve it at MS teams admin portal it's automatically reflecting me at both dev & test env app. Also, can't use .env file in spfx tab project.
**To Reproduce**
Steps to reproduce the behavior:
1. Create new spfx tab project from teamsfx cli.
2. Create new env's and provision them.
3. Deploy both env's app and approve at MS teams admin
4. Create new update for dev env and deploy it.
5. It reflects at both dev & test env apps.
**Expected behavior**
Dev update should not be reflect at other env if we deploy new update only at dev env.
**VS Code Extension Information (please complete the following information):**
- Teams Toolkit extension verion: 4.1.3
**CLI Information (please complete the following information):**
- Teamsfx-cli version: 1.2.4
|
1.0
|
TeamsFx spfx tab project Env's replicating each other - **Describe the bug**
Hi, I have created a spfx tab project from teamfx cli. i have created two env's dev & test which i am managing from teams toolkit exstension. when i deploy new update in dev env and approve it at MS teams admin portal it's automatically reflecting me at both dev & test env app. Also, can't use .env file in spfx tab project.
**To Reproduce**
Steps to reproduce the behavior:
1. Create new spfx tab project from teamsfx cli.
2. Create new env's and provision them.
3. Deploy both env's app and approve at MS teams admin
4. Create new update for dev env and deploy it.
5. It reflects at both dev & test env apps.
**Expected behavior**
Dev update should not be reflect at other env if we deploy new update only at dev env.
**VS Code Extension Information (please complete the following information):**
- Teams Toolkit extension verion: 4.1.3
**CLI Information (please complete the following information):**
- Teamsfx-cli version: 1.2.4
|
non_process
|
teamsfx spfx tab project env s replicating each other describe the bug hi i have created a spfx tab project from teamfx cli i have created two env s dev test which i am managing from teams toolkit exstension when i deploy new update in dev env and approve it at ms teams admin portal it s automatically reflecting me at both dev test env app also can t use env file in spfx tab project to reproduce steps to reproduce the behavior create new spfx tab project from teamsfx cli create new env s and provision them deploy both env s app and approve at ms teams admin create new update for dev env and deploy it it reflects at both dev test env apps expected behavior dev update should not be reflect at other env if we deploy new update only at dev env vs code extension information please complete the following information teams toolkit extension verion cli information please complete the following information teamsfx cli version
| 0
|
50,256
| 13,187,403,496
|
IssuesEvent
|
2020-08-13 03:18:23
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
closed
|
make_tarball_rootsys.sh.in bugfix (Trac #393)
|
Migrated from Trac combo core defect
|
cmake/make_tarball_rootsys.sh.in needs a recursive copy flag '-r' since new ROOT lib dir contains subdirectories.
Line 14:
- cp -P @ROOTSYS@/lib/* @CMAKE_INSTALL_PREFIX@/cernroot/lib
+ cp -r -P @ROOTSYS@/lib/* @CMAKE_INSTALL_PREFIX@/cernroot/lib
<details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/393
, reported by juancarlos and owned by blaufuss_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2012-06-03T04:44:16",
"description": "cmake/make_tarball_rootsys.sh.in needs a recursive copy flag '-r' since new ROOT lib dir contains subdirectories.\n\nLine 14:\n\n- cp -P @ROOTSYS@/lib/* @CMAKE_INSTALL_PREFIX@/cernroot/lib\n+ cp -r -P @ROOTSYS@/lib/* @CMAKE_INSTALL_PREFIX@/cernroot/lib",
"reporter": "juancarlos",
"cc": "",
"resolution": "fixed",
"_ts": "1338698656000000",
"component": "combo core",
"summary": "make_tarball_rootsys.sh.in bugfix",
"priority": "normal",
"keywords": "cmake",
"time": "2012-05-16T21:21:39",
"milestone": "",
"owner": "blaufuss",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
make_tarball_rootsys.sh.in bugfix (Trac #393) - cmake/make_tarball_rootsys.sh.in needs a recursive copy flag '-r' since new ROOT lib dir contains subdirectories.
Line 14:
- cp -P @ROOTSYS@/lib/* @CMAKE_INSTALL_PREFIX@/cernroot/lib
+ cp -r -P @ROOTSYS@/lib/* @CMAKE_INSTALL_PREFIX@/cernroot/lib
<details>
<summary>_Migrated from https://code.icecube.wisc.edu/ticket/393
, reported by juancarlos and owned by blaufuss_</summary>
<p>
```json
{
"status": "closed",
"changetime": "2012-06-03T04:44:16",
"description": "cmake/make_tarball_rootsys.sh.in needs a recursive copy flag '-r' since new ROOT lib dir contains subdirectories.\n\nLine 14:\n\n- cp -P @ROOTSYS@/lib/* @CMAKE_INSTALL_PREFIX@/cernroot/lib\n+ cp -r -P @ROOTSYS@/lib/* @CMAKE_INSTALL_PREFIX@/cernroot/lib",
"reporter": "juancarlos",
"cc": "",
"resolution": "fixed",
"_ts": "1338698656000000",
"component": "combo core",
"summary": "make_tarball_rootsys.sh.in bugfix",
"priority": "normal",
"keywords": "cmake",
"time": "2012-05-16T21:21:39",
"milestone": "",
"owner": "blaufuss",
"type": "defect"
}
```
</p>
</details>
|
non_process
|
make tarball rootsys sh in bugfix trac cmake make tarball rootsys sh in needs a recursive copy flag r since new root lib dir contains subdirectories line cp p rootsys lib cmake install prefix cernroot lib cp r p rootsys lib cmake install prefix cernroot lib migrated from reported by juancarlos and owned by blaufuss json status closed changetime description cmake make tarball rootsys sh in needs a recursive copy flag r since new root lib dir contains subdirectories n nline n n cp p rootsys lib cmake install prefix cernroot lib n cp r p rootsys lib cmake install prefix cernroot lib reporter juancarlos cc resolution fixed ts component combo core summary make tarball rootsys sh in bugfix priority normal keywords cmake time milestone owner blaufuss type defect
| 0
|
19,874
| 26,288,037,800
|
IssuesEvent
|
2023-01-08 03:17:44
|
fdhhhdjd/Web-Online-School-Libary-Book
|
https://api.github.com/repos/fdhhhdjd/Web-Online-School-Libary-Book
|
opened
|
Read File Excel ( Back-End )
|
DEV Processing
|
**** Read File Excel ****
- Library node-xlsx
- Code performance equal buffer.
- convert data JSON.
|
1.0
|
Read File Excel ( Back-End ) - **** Read File Excel ****
- Library node-xlsx
- Code performance equal buffer.
- convert data JSON.
|
process
|
read file excel back end read file excel library node xlsx code performance equal buffer convert data json
| 1
|
21,524
| 29,806,279,230
|
IssuesEvent
|
2023-06-16 11:56:57
|
parca-dev/parca-agent
|
https://api.github.com/repos/parca-dev/parca-agent
|
closed
|
Normalization issues
|
P0 area/process-mapping
|
i.e. while running `kubectl run -n parca debug --image=python:latest -it`
aka "the cockroachdb bug" (still present in `main`)
|
1.0
|
Normalization issues - i.e. while running `kubectl run -n parca debug --image=python:latest -it`
aka "the cockroachdb bug" (still present in `main`)
|
process
|
normalization issues i e while running kubectl run n parca debug image python latest it aka the cockroachdb bug still present in main
| 1
|
296,626
| 22,310,482,100
|
IssuesEvent
|
2022-06-13 16:31:04
|
minetest/minetest
|
https://api.github.com/repos/minetest/minetest
|
closed
|
moveresult.touching_ground value is incorrect or misleading
|
@ Documentation Question
|
##### Minetest version
<!--
Paste Minetest version between quotes below
If you are on a devel version, please add git commit hash
You can use `minetest --version` to find it.
-->
```
Minetest 5.3.0 (Linux)
Using Irrlicht 1.8.4
BUILD_TYPE=Release
RUN_IN_PLACE=0
USE_GETTEXT=1
USE_SOUND=1
USE_CURL=1
USE_FREETYPE=1
USE_LUAJIT=1
STATIC_SHAREDIR="/nix/store/lyl5b6q7rvv7pp6szn0krddy7102fnx6-minetest-5.3.0/share/minetest"
STATIC_LOCALEDIR="/nix/store/lyl5b6q7rvv7pp6szn0krddy7102fnx6-minetest-5.3.0/share/locale"
```
I've also tested with commit `4fcd000e20a26120349184cb9d40342b7876e6b8` from January 21st.
##### First of all
Thanks for all the hard work. I'm loving writing mods for minetest.
##### Summary
<!-- Describe your problem here -->
The collision info table received during the `on_step` function has a wrong or misleading value for `touching_ground`. This value only seems to be `true` if the entity has just moved. For instance, you can spawn a physical entity in the air and set it's velocity to negative y. It will fall and `touching_ground` will be `false`, which is correct. Once it touches the ground, `touching_ground` becomes `true`, which is also correct, and the velocity seems to be reset to 0 because of the collision. The problem is that on the next `on_step`, `touching_ground` is now `false`, even though the entity is literally on the ground.
My guess is that collisions were not calculated because there is no movement, but somehow the engine didn't account for the fact that the entity is *already* on the ground.
This seems like a weird behavior to me, and I couldn't find a good way to determine that an entity is on the ground from Lua. Am I missing something here? If there's a good solution for this, I'm happy to write documentation about it.
|
1.0
|
moveresult.touching_ground value is incorrect or misleading - ##### Minetest version
<!--
Paste Minetest version between quotes below
If you are on a devel version, please add git commit hash
You can use `minetest --version` to find it.
-->
```
Minetest 5.3.0 (Linux)
Using Irrlicht 1.8.4
BUILD_TYPE=Release
RUN_IN_PLACE=0
USE_GETTEXT=1
USE_SOUND=1
USE_CURL=1
USE_FREETYPE=1
USE_LUAJIT=1
STATIC_SHAREDIR="/nix/store/lyl5b6q7rvv7pp6szn0krddy7102fnx6-minetest-5.3.0/share/minetest"
STATIC_LOCALEDIR="/nix/store/lyl5b6q7rvv7pp6szn0krddy7102fnx6-minetest-5.3.0/share/locale"
```
I've also tested with commit `4fcd000e20a26120349184cb9d40342b7876e6b8` from January 21st.
##### First of all
Thanks for all the hard work. I'm loving writing mods for minetest.
##### Summary
<!-- Describe your problem here -->
The collision info table received during the `on_step` function has a wrong or misleading value for `touching_ground`. This value only seems to be `true` if the entity has just moved. For instance, you can spawn a physical entity in the air and set it's velocity to negative y. It will fall and `touching_ground` will be `false`, which is correct. Once it touches the ground, `touching_ground` becomes `true`, which is also correct, and the velocity seems to be reset to 0 because of the collision. The problem is that on the next `on_step`, `touching_ground` is now `false`, even though the entity is literally on the ground.
My guess is that collisions were not calculated because there is no movement, but somehow the engine didn't account for the fact that the entity is *already* on the ground.
This seems like a weird behavior to me, and I couldn't find a good way to determine that an entity is on the ground from Lua. Am I missing something here? If there's a good solution for this, I'm happy to write documentation about it.
|
non_process
|
moveresult touching ground value is incorrect or misleading minetest version paste minetest version between quotes below if you are on a devel version please add git commit hash you can use minetest version to find it minetest linux using irrlicht build type release run in place use gettext use sound use curl use freetype use luajit static sharedir nix store minetest share minetest static localedir nix store minetest share locale i ve also tested with commit from january first of all thanks for all the hard work i m loving writing mods for minetest summary the collision info table received during the on step function has a wrong or misleading value for touching ground this value only seems to be true if the entity has just moved for instance you can spawn a physical entity in the air and set it s velocity to negative y it will fall and touching ground will be false which is correct once it touches the ground touching ground becomes true which is also correct and the velocity seems to be reset to because of the collision the problem is that on the next on step touching ground is now false even though the entity is literally on the ground my guess is that collisions were not calculated because there is no movement but somehow the engine didn t account for the fact that the entity is already on the ground this seems like a weird behavior to me and i couldn t find a good way to determine that an entity is on the ground from lua am i missing something here if there s a good solution for this i m happy to write documentation about it
| 0
|
6,508
| 6,490,278,376
|
IssuesEvent
|
2017-08-21 06:38:15
|
camptocamp/c2cgeoportal
|
https://api.github.com/repos/camptocamp/c2cgeoportal
|
closed
|
Suspicious relativ path to build the project standalone
|
Infrastructure Ready
|
Appears on c2cgeoportal master (20 feb 2017)
I've tried to to a make build on the project but these two lines make issues:
- https://github.com/camptocamp/c2cgeoportal/blob/daca42c1b3469531b7dc176ed3ffa5d368ede658/Makefile#L1 (`/build` shouldn't be `build` ?)
- https://github.com/camptocamp/c2cgeoportal/blob/daca42c1b3469531b7dc176ed3ffa5d368ede658/Makefile#L61 ?
Absolute path... that's not false ?
|
1.0
|
Suspicious relativ path to build the project standalone - Appears on c2cgeoportal master (20 feb 2017)
I've tried to to a make build on the project but these two lines make issues:
- https://github.com/camptocamp/c2cgeoportal/blob/daca42c1b3469531b7dc176ed3ffa5d368ede658/Makefile#L1 (`/build` shouldn't be `build` ?)
- https://github.com/camptocamp/c2cgeoportal/blob/daca42c1b3469531b7dc176ed3ffa5d368ede658/Makefile#L61 ?
Absolute path... that's not false ?
|
non_process
|
suspicious relativ path to build the project standalone appears on master feb i ve tried to to a make build on the project but these two lines make issues build shouldn t be build absolute path that s not false
| 0
|
411,031
| 27,811,102,130
|
IssuesEvent
|
2023-03-18 05:33:39
|
Real-Dev-Squad/website-api-contracts
|
https://api.github.com/repos/Real-Dev-Squad/website-api-contracts
|
closed
|
API Contract for GET Idle users/members
|
documentation
|
### AC
- Create an API contract for `/users/idle` and `/users/idle?members=true`
### Link with
Issue [#635](https://github.com/Real-Dev-Squad/website-backend/issues/635)
|
1.0
|
API Contract for GET Idle users/members - ### AC
- Create an API contract for `/users/idle` and `/users/idle?members=true`
### Link with
Issue [#635](https://github.com/Real-Dev-Squad/website-backend/issues/635)
|
non_process
|
api contract for get idle users members ac create an api contract for users idle and users idle members true link with issue
| 0
|
163
| 2,583,797,330
|
IssuesEvent
|
2015-02-16 10:17:32
|
luc-github/Repetier-Firmware-0.92
|
https://api.github.com/repos/luc-github/Repetier-Firmware-0.92
|
closed
|
Build fails on Arduino Nightly (1.6.0) Jan 14
|
enhancement Waiting to be processed
|
It looks like the current nightly builds of Arduino have incompatible changes with variants.cpp. I was able to build by copying the "hardware" directory from 1.5.8.
|
1.0
|
Build fails on Arduino Nightly (1.6.0) Jan 14 - It looks like the current nightly builds of Arduino have incompatible changes with variants.cpp. I was able to build by copying the "hardware" directory from 1.5.8.
|
process
|
build fails on arduino nightly jan it looks like the current nightly builds of arduino have incompatible changes with variants cpp i was able to build by copying the hardware directory from
| 1
|
4,394
| 7,285,884,163
|
IssuesEvent
|
2018-02-23 07:12:33
|
muflihun/residue
|
https://api.github.com/repos/muflihun/residue
|
closed
|
Remove plain log request support
|
area: log-processing type: improvement
|
This is a security concern + extra instructions for potentially unused feature
We will remove it in 1.5.0
|
1.0
|
Remove plain log request support - This is a security concern + extra instructions for potentially unused feature
We will remove it in 1.5.0
|
process
|
remove plain log request support this is a security concern extra instructions for potentially unused feature we will remove it in
| 1
|
115,529
| 14,799,011,046
|
IssuesEvent
|
2021-01-13 01:13:01
|
vmware-tanzu/antrea
|
https://api.github.com/repos/vmware-tanzu/antrea
|
opened
|
Replace hack/netpol/ with new upstream NetworkPolicy test suite
|
area/test/community kind/design priority/important-longterm
|
The netpol test suite that we protoyped in Antrea (https://github.com/vmware-tanzu/antrea/tree/master/hack/netpol) was ported upstream by @jayunit100 and others: https://github.com/kubernetes/kubernetes/tree/master/test/e2e/network/netpol
As a result it now makes sense to remove the `hack/netpol/` directory altogether and instead start running the upstream version of the test suite as part of Antrea CI. I think we can probably move from a Kind CI job to a Jenkins job on VMC (VMware on AWS), to avoid some known issues with Kind / the Open vSwitch netdev datapath (see https://github.com/vmware-tanzu/antrea/issues/897 for an example).
My preference would be to simply run it as part of the existing `jenkins-networkpolicy` job:
* because it makes sense :)
* to avoid introducing yet another job
* because the test suite is supposed to run pretty fast and should not add too much time compared to the current job
I am hoping we can simply update https://github.com/vmware-tanzu/antrea/blob/master/ci/run-k8s-e2e-tests.sh to avoid increasing the number of CI scripts we have to maintain.
The only issue I see is that at this time (01/12), there is no named tag of the `k8s.gcr.io/conformance` image which includes the upstream netpol tests. The latest tag seems to be `v1.21.0-alpha.0`, and it doesn't include them.
As a result, we can either:
1. wait for a version of the `k8s.gcr.io/conformance` image with support for the netpol tests
2. build the `k8s.gcr.io/conformance` image from the K8s source (https://github.com/kubernetes/kubernetes/blob/master/cluster/images/conformance/Makefile) and use it in `run-k8s-e2e-tests.sh`
I don't really have a preference for either. If someone wants to start working on this issue in the near future, then they can go with the second solution.
|
1.0
|
Replace hack/netpol/ with new upstream NetworkPolicy test suite - The netpol test suite that we protoyped in Antrea (https://github.com/vmware-tanzu/antrea/tree/master/hack/netpol) was ported upstream by @jayunit100 and others: https://github.com/kubernetes/kubernetes/tree/master/test/e2e/network/netpol
As a result it now makes sense to remove the `hack/netpol/` directory altogether and instead start running the upstream version of the test suite as part of Antrea CI. I think we can probably move from a Kind CI job to a Jenkins job on VMC (VMware on AWS), to avoid some known issues with Kind / the Open vSwitch netdev datapath (see https://github.com/vmware-tanzu/antrea/issues/897 for an example).
My preference would be to simply run it as part of the existing `jenkins-networkpolicy` job:
* because it makes sense :)
* to avoid introducing yet another job
* because the test suite is supposed to run pretty fast and should not add too much time compared to the current job
I am hoping we can simply update https://github.com/vmware-tanzu/antrea/blob/master/ci/run-k8s-e2e-tests.sh to avoid increasing the number of CI scripts we have to maintain.
The only issue I see is that at this time (01/12), there is no named tag of the `k8s.gcr.io/conformance` image which includes the upstream netpol tests. The latest tag seems to be `v1.21.0-alpha.0`, and it doesn't include them.
As a result, we can either:
1. wait for a version of the `k8s.gcr.io/conformance` image with support for the netpol tests
2. build the `k8s.gcr.io/conformance` image from the K8s source (https://github.com/kubernetes/kubernetes/blob/master/cluster/images/conformance/Makefile) and use it in `run-k8s-e2e-tests.sh`
I don't really have a preference for either. If someone wants to start working on this issue in the near future, then they can go with the second solution.
|
non_process
|
replace hack netpol with new upstream networkpolicy test suite the netpol test suite that we protoyped in antrea was ported upstream by and others as a result it now makes sense to remove the hack netpol directory altogether and instead start running the upstream version of the test suite as part of antrea ci i think we can probably move from a kind ci job to a jenkins job on vmc vmware on aws to avoid some known issues with kind the open vswitch netdev datapath see for an example my preference would be to simply run it as part of the existing jenkins networkpolicy job because it makes sense to avoid introducing yet another job because the test suite is supposed to run pretty fast and should not add too much time compared to the current job i am hoping we can simply update to avoid increasing the number of ci scripts we have to maintain the only issue i see is that at this time there is no named tag of the gcr io conformance image which includes the upstream netpol tests the latest tag seems to be alpha and it doesn t include them as a result we can either wait for a version of the gcr io conformance image with support for the netpol tests build the gcr io conformance image from the source and use it in run tests sh i don t really have a preference for either if someone wants to start working on this issue in the near future then they can go with the second solution
| 0
|
615,216
| 19,250,014,456
|
IssuesEvent
|
2021-12-09 03:17:19
|
matrixorigin/matrixone
|
https://api.github.com/repos/matrixorigin/matrixone
|
opened
|
add AOE RFC documents
|
component/aoe priority/high kind/feature severity/critical
|
1. Overall architecture
2. WAL
3. Buffer manager
4. Metadata
5. Data and index management
6. MVCC
7. Logstore
|
1.0
|
add AOE RFC documents - 1. Overall architecture
2. WAL
3. Buffer manager
4. Metadata
5. Data and index management
6. MVCC
7. Logstore
|
non_process
|
add aoe rfc documents overall architecture wal buffer manager metadata data and index management mvcc logstore
| 0
|
412,455
| 27,859,147,553
|
IssuesEvent
|
2023-03-21 03:34:54
|
CodeforHawaii/HIERR
|
https://api.github.com/repos/CodeforHawaii/HIERR
|
closed
|
Define user stories
|
documentation High Priority
|
This is for defining what user stories there are. Once this list is defined, we will send to Scott for prioritization.
|
1.0
|
Define user stories - This is for defining what user stories there are. Once this list is defined, we will send to Scott for prioritization.
|
non_process
|
define user stories this is for defining what user stories there are once this list is defined we will send to scott for prioritization
| 0
|
372,020
| 11,007,923,144
|
IssuesEvent
|
2019-12-04 09:31:36
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
t.co - see bug description
|
browser-focus-geckoview engine-gecko priority-critical
|
<!-- @browser: Firefox Mobile 71.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:71.0) Gecko/71.0 Firefox/71.0 -->
<!-- @reported_with: -->
<!-- @extra_labels: browser-focus-geckoview -->
**URL**: https://t.co/YDbDt5izbJ?amp=1
**Browser / Version**: Firefox Mobile 71.0
**Operating System**: Android
**Tested Another Browser**: Yes
**Problem type**: Something else
**Description**: Firefox focus does not trust user level https certificates, making Adguard https filtering broken
**Steps to Reproduce**:
Firefox focus does not trust user level https certificates, making Adguard https filtering broken
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
t.co - see bug description - <!-- @browser: Firefox Mobile 71.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:71.0) Gecko/71.0 Firefox/71.0 -->
<!-- @reported_with: -->
<!-- @extra_labels: browser-focus-geckoview -->
**URL**: https://t.co/YDbDt5izbJ?amp=1
**Browser / Version**: Firefox Mobile 71.0
**Operating System**: Android
**Tested Another Browser**: Yes
**Problem type**: Something else
**Description**: Firefox focus does not trust user level https certificates, making Adguard https filtering broken
**Steps to Reproduce**:
Firefox focus does not trust user level https certificates, making Adguard https filtering broken
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
t co see bug description url browser version firefox mobile operating system android tested another browser yes problem type something else description firefox focus does not trust user level https certificates making adguard https filtering broken steps to reproduce firefox focus does not trust user level https certificates making adguard https filtering broken browser configuration none from with ❤️
| 0
|
10,448
| 13,227,205,896
|
IssuesEvent
|
2020-08-18 02:25:35
|
googlemaps/v3-utility-library
|
https://api.github.com/repos/googlemaps/v3-utility-library
|
closed
|
markerclustererplus unit tests
|
help wanted priority: p3 stale type: process
|
recently had issues such as #549 that could have been caught through unittests
|
1.0
|
markerclustererplus unit tests - recently had issues such as #549 that could have been caught through unittests
|
process
|
markerclustererplus unit tests recently had issues such as that could have been caught through unittests
| 1
|
191
| 2,519,739,944
|
IssuesEvent
|
2015-01-18 09:00:56
|
mbunkus/mtx-trac-import-test
|
https://api.github.com/repos/mbunkus/mtx-trac-import-test
|
opened
|
mkvmerge will freeze if used with many tracks
|
C: mkvmerge P: normal R: fixed T: defect
|
**Reported by moritz on 9 Aug 2003 20:40 UTC**
Liisachan reports:
I tested 3 samples made by MatroskaMuxer
(1) Xvid + 1 Vorbis + 5 ssa = 7 tracks
(2) xvid + 1 vorbis + 16 ssa = 18 tracks
(3) xvid + 1 vorbis + 17 ssa = 19 tracks
Command line is like:
mkvmerge -o "out.mkv" --language 1:rus --language 2:rus
--language 3:eng --language 4:jpn --language 5:fre --language
6:dut --language 7:rus "in.mkv" > debug.txt
mkvmerge 0.5.0 works fine for the all,
while 0.6.0 works only for (1)
0.6.0 will freeze for (2) and (3)
0.6.0 can handle Chapter / Attachment files for (1) but 0.6.0
cannot handle (2) and (3) even without chapters etc...
0.6.0 doesnt't like too many tracks, i wonder?
Migrated-From: https://trac.bunkus.org/ticket/8
|
1.0
|
mkvmerge will freeze if used with many tracks - **Reported by moritz on 9 Aug 2003 20:40 UTC**
Liisachan reports:
I tested 3 samples made by MatroskaMuxer
(1) Xvid + 1 Vorbis + 5 ssa = 7 tracks
(2) xvid + 1 vorbis + 16 ssa = 18 tracks
(3) xvid + 1 vorbis + 17 ssa = 19 tracks
Command line is like:
mkvmerge -o "out.mkv" --language 1:rus --language 2:rus
--language 3:eng --language 4:jpn --language 5:fre --language
6:dut --language 7:rus "in.mkv" > debug.txt
mkvmerge 0.5.0 works fine for the all,
while 0.6.0 works only for (1)
0.6.0 will freeze for (2) and (3)
0.6.0 can handle Chapter / Attachment files for (1) but 0.6.0
cannot handle (2) and (3) even without chapters etc...
0.6.0 doesnt't like too many tracks, i wonder?
Migrated-From: https://trac.bunkus.org/ticket/8
|
non_process
|
mkvmerge will freeze if used with many tracks reported by moritz on aug utc liisachan reports i tested samples made by matroskamuxer xvid vorbis ssa tracks xvid vorbis ssa tracks xvid vorbis ssa tracks command line is like mkvmerge o out mkv language rus language rus language eng language jpn language fre language dut language rus in mkv debug txt mkvmerge works fine for the all while works only for will freeze for and can handle chapter attachment files for but cannot handle and even without chapters etc doesnt t like too many tracks i wonder migrated from
| 0
|
12,839
| 15,223,150,502
|
IssuesEvent
|
2021-02-18 01:58:51
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Allow tool output and export to save a new layer/table in existing databases also for non QGIS native tools.
|
Feature Request Feedback Processing stale
|
First, it'd be nice to be able to save the output of a tool directly to a PostGIS layer/table. It seems that the only way to do this right now is to dump it out to a different format (shapefile/geojson) and then import it to PostGIS.
Second, it would be nice to save tool output as a new layer in an existing database. Many tools can save layers to a **new** geopackage or spatialite database, but it doesn't seem possible to save the output as a new layer/table in an **existing** database. I like keeping all of my layers for a project in a single database - that makes it easier to move around. And I also use SQL in a lot of my analyses -it's fairly clunky to do joins between tables in separate databases.
Thanks to all of the QGIS developers out there! I'm more and more impressed with every release!
|
1.0
|
Allow tool output and export to save a new layer/table in existing databases also for non QGIS native tools. - First, it'd be nice to be able to save the output of a tool directly to a PostGIS layer/table. It seems that the only way to do this right now is to dump it out to a different format (shapefile/geojson) and then import it to PostGIS.
Second, it would be nice to save tool output as a new layer in an existing database. Many tools can save layers to a **new** geopackage or spatialite database, but it doesn't seem possible to save the output as a new layer/table in an **existing** database. I like keeping all of my layers for a project in a single database - that makes it easier to move around. And I also use SQL in a lot of my analyses -it's fairly clunky to do joins between tables in separate databases.
Thanks to all of the QGIS developers out there! I'm more and more impressed with every release!
|
process
|
allow tool output and export to save a new layer table in existing databases also for non qgis native tools first it d be nice to be able to save the output of a tool directly to a postgis layer table it seems that the only way to do this right now is to dump it out to a different format shapefile geojson and then import it to postgis second it would be nice to save tool output as a new layer in an existing database many tools can save layers to a new geopackage or spatialite database but it doesn t seem possible to save the output as a new layer table in an existing database i like keeping all of my layers for a project in a single database that makes it easier to move around and i also use sql in a lot of my analyses it s fairly clunky to do joins between tables in separate databases thanks to all of the qgis developers out there i m more and more impressed with every release
| 1
|
18,119
| 24,150,759,854
|
IssuesEvent
|
2022-09-22 00:17:44
|
allinurl/goaccess
|
https://api.github.com/repos/allinurl/goaccess
|
closed
|
Parse Log with hostname instead of ipv4/ipv6
|
question log-processing command-line options
|
I have a log-file where the ip is already resolved and instead of ip has a hostname as a field.
It seems this field is not understood by goaccess. As I understand the hostname field is manadatory, but is it really a requirement to hold an IP?
`Token 'some.perfect-valid-hostname.example.com' doesn't match specifier '%h'`
Does someone have an advice?
|
1.0
|
Parse Log with hostname instead of ipv4/ipv6 - I have a log-file where the ip is already resolved and instead of ip has a hostname as a field.
It seems this field is not understood by goaccess. As I understand the hostname field is manadatory, but is it really a requirement to hold an IP?
`Token 'some.perfect-valid-hostname.example.com' doesn't match specifier '%h'`
Does someone have an advice?
|
process
|
parse log with hostname instead of i have a log file where the ip is already resolved and instead of ip has a hostname as a field it seems this field is not understood by goaccess as i understand the hostname field is manadatory but is it really a requirement to hold an ip token some perfect valid hostname example com doesn t match specifier h does someone have an advice
| 1
|
42,870
| 7,006,737,191
|
IssuesEvent
|
2017-12-19 09:39:54
|
rkt/rkt
|
https://api.github.com/repos/rkt/rkt
|
closed
|
Describe how to run unit/functional tests in the hacking guide.
|
area/developer tooling kind/cleanup kind/documentation
|
To make new developers' life easier, we need to document/link the guide on how to run unit/functional tests locally.
|
1.0
|
Describe how to run unit/functional tests in the hacking guide. - To make new developers' life easier, we need to document/link the guide on how to run unit/functional tests locally.
|
non_process
|
describe how to run unit functional tests in the hacking guide to make new developers life easier we need to document link the guide on how to run unit functional tests locally
| 0
|
5,304
| 8,124,297,086
|
IssuesEvent
|
2018-08-16 17:05:12
|
dotnet/corefx
|
https://api.github.com/repos/dotnet/corefx
|
closed
|
Intermittent: Empty Process stdout
|
area-System.Diagnostics.Process needs more info question
|
## Story
I don't know why, but sometimes when running same console program with same input data, System.Diagnostics fucntionality returns empty stdout.
It's not a problem with running process. Also I checked my code many times, no errors found.
## Platforms
- Windows (tested)
- Linux (tested)
## I think this non-repro code would be useful for creation of working repro of this problem
I tried to do a minimal repro, but all tests passed and none failed... so maybe something wrong with my code. Or, if someone have same problem, maybe something wrong in .NET Core.
````c#
using System;
using System.IO;
using System.Text;
using System.Threading;
using System.Diagnostics;
class EmptyStdOutRepro
{
public static void Main()
{
new EmptyStdOutRepro().Run();
}
public void Run()
{
while (true)
Exec();
}
private void Exec()
{
int lineCount = 0;
string output = "";
Process process = new Process();
process.StartInfo.FileName = "ipconfig.exe";
process.StartInfo.UseShellExecute = false;
process.StartInfo.RedirectStandardOutput = true;
process.StartInfo.CreateNoWindow = true;
process.StartInfo.ErrorDialog = false;
process.OutputDataReceived += new DataReceivedEventHandler((sender, e) =>
{
if (!String.IsNullOrEmpty(e.Data))
{
lineCount++;
output += ("\n[" + lineCount + "]: " + e.Data);
}
});
process.Start();
process.BeginOutputReadLine();
process.WaitForExit();
if (output != "")
Console.WriteLine("PASSED!" + output.Length);
else
Console.WriteLine("FAILED!" + output.Length);
process.WaitForExit();
process.Close();
}
}
````
## Working repro, needs minimization
Class ProgramExecutor in SimplePM_Server repo: https://github.com/SirkadirovTeam/SimplePM_Server/tree/master/SimplePM_Server/ProgramTesting/SRunner/ProgramExecutor
|
1.0
|
Intermittent: Empty Process stdout - ## Story
I don't know why, but sometimes when running same console program with same input data, System.Diagnostics fucntionality returns empty stdout.
It's not a problem with running process. Also I checked my code many times, no errors found.
## Platforms
- Windows (tested)
- Linux (tested)
## I think this non-repro code would be useful for creation of working repro of this problem
I tried to do a minimal repro, but all tests passed and none failed... so maybe something wrong with my code. Or, if someone have same problem, maybe something wrong in .NET Core.
````c#
using System;
using System.IO;
using System.Text;
using System.Threading;
using System.Diagnostics;
class EmptyStdOutRepro
{
public static void Main()
{
new EmptyStdOutRepro().Run();
}
public void Run()
{
while (true)
Exec();
}
private void Exec()
{
int lineCount = 0;
string output = "";
Process process = new Process();
process.StartInfo.FileName = "ipconfig.exe";
process.StartInfo.UseShellExecute = false;
process.StartInfo.RedirectStandardOutput = true;
process.StartInfo.CreateNoWindow = true;
process.StartInfo.ErrorDialog = false;
process.OutputDataReceived += new DataReceivedEventHandler((sender, e) =>
{
if (!String.IsNullOrEmpty(e.Data))
{
lineCount++;
output += ("\n[" + lineCount + "]: " + e.Data);
}
});
process.Start();
process.BeginOutputReadLine();
process.WaitForExit();
if (output != "")
Console.WriteLine("PASSED!" + output.Length);
else
Console.WriteLine("FAILED!" + output.Length);
process.WaitForExit();
process.Close();
}
}
````
## Working repro, needs minimization
Class ProgramExecutor in SimplePM_Server repo: https://github.com/SirkadirovTeam/SimplePM_Server/tree/master/SimplePM_Server/ProgramTesting/SRunner/ProgramExecutor
|
process
|
intermittent empty process stdout story i don t know why but sometimes when running same console program with same input data system diagnostics fucntionality returns empty stdout it s not a problem with running process also i checked my code many times no errors found platforms windows tested linux tested i think this non repro code would be useful for creation of working repro of this problem i tried to do a minimal repro but all tests passed and none failed so maybe something wrong with my code or if someone have same problem maybe something wrong in net core c using system using system io using system text using system threading using system diagnostics class emptystdoutrepro public static void main new emptystdoutrepro run public void run while true exec private void exec int linecount string output process process new process process startinfo filename ipconfig exe process startinfo useshellexecute false process startinfo redirectstandardoutput true process startinfo createnowindow true process startinfo errordialog false process outputdatareceived new datareceivedeventhandler sender e if string isnullorempty e data linecount output n e data process start process beginoutputreadline process waitforexit if output console writeline passed output length else console writeline failed output length process waitforexit process close working repro needs minimization class programexecutor in simplepm server repo
| 1
|
331,247
| 28,760,770,983
|
IssuesEvent
|
2023-05-01 00:03:24
|
WordPress/gutenberg
|
https://api.github.com/repos/WordPress/gutenberg
|
closed
|
[Flaky Test] User Mention: should allow newlines after multiple mention completions
|
[Status] Stale [Type] Flaky Test
|
<!-- __META_DATA__:{} -->
**Flaky test detected. This is an auto-generated issue by GitHub Actions. Please do NOT edit this manually.**
## Test title
User Mention: should allow newlines after multiple mention completions
## Test path
`/test/e2e/specs/editor/various/autocomplete-and-mentions.spec.js`
## Errors
<!-- __TEST_RESULTS_LIST__ -->
<!-- __TEST_RESULT__ --><time datetime="2023-02-27T19:57:03.719Z"><code>[2023-02-27T19:57:03.719Z]</code></time> Test passed after 5 failed attempts on <a href="https://github.com/WordPress/gutenberg/actions/runs/4285345332"><code>try/set-attrs-idle</code></a>.<!-- /__TEST_RESULT__ -->
<br/>
<!-- __TEST_RESULT__ --><details>
<summary>
<time datetime="2023-03-30T07:42:51.930Z"><code>[2023-03-30T07:42:51.930Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/4561612597"><code>fix/padding-input-in-global-styles</code></a>.
</summary>
```
Error: keyboard.type: Target page, context or browser has been closed
at /home/runner/work/gutenberg/gutenberg/test/e2e/specs/editor/various/autocomplete-and-mentions.spec.js:403:25
```
</details><!-- /__TEST_RESULT__ -->
<!-- /__TEST_RESULTS_LIST__ -->
|
1.0
|
[Flaky Test] User Mention: should allow newlines after multiple mention completions - <!-- __META_DATA__:{} -->
**Flaky test detected. This is an auto-generated issue by GitHub Actions. Please do NOT edit this manually.**
## Test title
User Mention: should allow newlines after multiple mention completions
## Test path
`/test/e2e/specs/editor/various/autocomplete-and-mentions.spec.js`
## Errors
<!-- __TEST_RESULTS_LIST__ -->
<!-- __TEST_RESULT__ --><time datetime="2023-02-27T19:57:03.719Z"><code>[2023-02-27T19:57:03.719Z]</code></time> Test passed after 5 failed attempts on <a href="https://github.com/WordPress/gutenberg/actions/runs/4285345332"><code>try/set-attrs-idle</code></a>.<!-- /__TEST_RESULT__ -->
<br/>
<!-- __TEST_RESULT__ --><details>
<summary>
<time datetime="2023-03-30T07:42:51.930Z"><code>[2023-03-30T07:42:51.930Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/4561612597"><code>fix/padding-input-in-global-styles</code></a>.
</summary>
```
Error: keyboard.type: Target page, context or browser has been closed
at /home/runner/work/gutenberg/gutenberg/test/e2e/specs/editor/various/autocomplete-and-mentions.spec.js:403:25
```
</details><!-- /__TEST_RESULT__ -->
<!-- /__TEST_RESULTS_LIST__ -->
|
non_process
|
user mention should allow newlines after multiple mention completions flaky test detected this is an auto generated issue by github actions please do not edit this manually test title user mention should allow newlines after multiple mention completions test path test specs editor various autocomplete and mentions spec js errors test passed after failed attempts on test passed after failed attempt on a href error keyboard type target page context or browser has been closed at home runner work gutenberg gutenberg test specs editor various autocomplete and mentions spec js
| 0
|
22,325
| 30,913,247,439
|
IssuesEvent
|
2023-08-05 01:26:45
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
child_process,tty: pseudo-tty support
|
child_process feature request libuv blocked stale
|
Blocked on someone implementing https://github.com/libuv/libuv/issues/2640, then exposing it in node.
Refs #31409 and others.
|
1.0
|
child_process,tty: pseudo-tty support - Blocked on someone implementing https://github.com/libuv/libuv/issues/2640, then exposing it in node.
Refs #31409 and others.
|
process
|
child process tty pseudo tty support blocked on someone implementing then exposing it in node refs and others
| 1
|
111,587
| 9,535,889,137
|
IssuesEvent
|
2019-04-30 08:17:56
|
chameleon-system/chameleon-system
|
https://api.github.com/repos/chameleon-system/chameleon-system
|
closed
|
Image Sharpening fails on small images with image magick 6.9.x
|
Priority: High Status: Test Type: Bug
|
**Describe the bug**
When sharpening images is enabled for all image sizes, then thumbnail creation will fail for smaller images (example image size tested was 59x26)
**Affected version(s)**
Affects all chameleon versions
**To Reproduce**
Steps to reproduce the behavior:
1. In the cms backend go to cms settings and open the tab images
2. Add an entry without a size limitation and sharpening enabled
3. remove all thumbnails and clear cache
4. attempt to view the image via frontend
**Expected behavior**
A sharpend version of the image is generated. Instead, an error will be show in dev mode. in prod mode, no error message is shown but the image is not generated.
**Technical details**
source of the problem appears to be the parameter `-unsharp 1.5×1.0+1.5+0.02` being added to the imagemagick command in `\imageMagick::ResizeImage`. Executing the same command directly on the command line will produce the following error:
```
convert: invalid argument for option `-unsharp': 1.5×1.0+1.5+0.02 @ error/convert.c/ConvertImageCommand/3128.
```
The error occurs with imagemagick 6.9.7-4 Q16 x86_64 2017011 but not with version 6.8.9-9 Q16 x86_64 2018-11-11
Replacing `-unsharp 1.5×1.0+1.5+0.02` with `-adaptive-sharpen 0x1` appears to produce similar sharpness results and works for small images. This may be an alternative.
|
1.0
|
Image Sharpening fails on small images with image magick 6.9.x - **Describe the bug**
When sharpening images is enabled for all image sizes, then thumbnail creation will fail for smaller images (example image size tested was 59x26)
**Affected version(s)**
Affects all chameleon versions
**To Reproduce**
Steps to reproduce the behavior:
1. In the cms backend go to cms settings and open the tab images
2. Add an entry without a size limitation and sharpening enabled
3. remove all thumbnails and clear cache
4. attempt to view the image via frontend
**Expected behavior**
A sharpend version of the image is generated. Instead, an error will be show in dev mode. in prod mode, no error message is shown but the image is not generated.
**Technical details**
source of the problem appears to be the parameter `-unsharp 1.5×1.0+1.5+0.02` being added to the imagemagick command in `\imageMagick::ResizeImage`. Executing the same command directly on the command line will produce the following error:
```
convert: invalid argument for option `-unsharp': 1.5×1.0+1.5+0.02 @ error/convert.c/ConvertImageCommand/3128.
```
The error occurs with imagemagick 6.9.7-4 Q16 x86_64 2017011 but not with version 6.8.9-9 Q16 x86_64 2018-11-11
Replacing `-unsharp 1.5×1.0+1.5+0.02` with `-adaptive-sharpen 0x1` appears to produce similar sharpness results and works for small images. This may be an alternative.
|
non_process
|
image sharpening fails on small images with image magick x describe the bug when sharpening images is enabled for all image sizes then thumbnail creation will fail for smaller images example image size tested was affected version s affects all chameleon versions to reproduce steps to reproduce the behavior in the cms backend go to cms settings and open the tab images add an entry without a size limitation and sharpening enabled remove all thumbnails and clear cache attempt to view the image via frontend expected behavior a sharpend version of the image is generated instead an error will be show in dev mode in prod mode no error message is shown but the image is not generated technical details source of the problem appears to be the parameter unsharp × being added to the imagemagick command in imagemagick resizeimage executing the same command directly on the command line will produce the following error convert invalid argument for option unsharp × error convert c convertimagecommand the error occurs with imagemagick but not with version replacing unsharp × with adaptive sharpen appears to produce similar sharpness results and works for small images this may be an alternative
| 0
|
17,575
| 23,387,109,980
|
IssuesEvent
|
2022-08-11 14:35:57
|
vectordotdev/vector
|
https://api.github.com/repos/vectordotdev/vector
|
closed
|
New `recollect_fields` transform
|
needs: approval domain: transforms type: feature domain: processing
|
It would be nice to have a transform that, if a given field is not present in the event,
can insert this field with last observed value of this field from previous events.
Example configuration:
```toml
[transforms.recollect_fields]
inputs = ["in"]
fields = ["field1", "field2"]
```
Such a transform could have been implemented by a user with a stateful JavaScript transform (#721) as
```javascript
const fields = ["field1", "field2"];
const storedValues = {};
const handler = event => {
for (const field of fields) {
if (event.hasOwnProperty(field)) {
storedValues[field] = event[field];
} else if (storedValues.hasOwnProperty(field)) {
event[field] = storedValues[field];
}
}
return event;
}
```
As a bonus, this transform would allow to join multiple streams of events.
For example, for this config
```toml
[inputs.cpu_temp]
# produces events with field "cpu_temp"
[inputs.cpu_freq]
# produces events with field "cpu_freq"
[transforms.cpu_state]
type = "recollect_fields"
inputs = ["cpu_temp", "cpu_freq"]
fields = ["cpu_temp", "cpu_freq"]
[outputs.console]
type = "console"
inputs = ["recollect_fields"]
```
events containing both temperature and frequency would be printed.
In addition to the most basic form described above, we can add an optional configuration parameter `forget_time` which would prevent the field value from being recollected if it was produced earlier than the the current time minus timeout.
|
1.0
|
New `recollect_fields` transform - It would be nice to have a transform that, if a given field is not present in the event,
can insert this field with last observed value of this field from previous events.
Example configuration:
```toml
[transforms.recollect_fields]
inputs = ["in"]
fields = ["field1", "field2"]
```
Such a transform could have been implemented by a user with a stateful JavaScript transform (#721) as
```javascript
const fields = ["field1", "field2"];
const storedValues = {};
const handler = event => {
for (const field of fields) {
if (event.hasOwnProperty(field)) {
storedValues[field] = event[field];
} else if (storedValues.hasOwnProperty(field)) {
event[field] = storedValues[field];
}
}
return event;
}
```
As a bonus, this transform would allow to join multiple streams of events.
For example, for this config
```toml
[inputs.cpu_temp]
# produces events with field "cpu_temp"
[inputs.cpu_freq]
# produces events with field "cpu_freq"
[transforms.cpu_state]
type = "recollect_fields"
inputs = ["cpu_temp", "cpu_freq"]
fields = ["cpu_temp", "cpu_freq"]
[outputs.console]
type = "console"
inputs = ["recollect_fields"]
```
events containing both temperature and frequency would be printed.
In addition to the most basic form described above, we can add an optional configuration parameter `forget_time` which would prevent the field value from being recollected if it was produced earlier than the the current time minus timeout.
|
process
|
new recollect fields transform it would be nice to have a transform that if a given field is not present in the event can insert this field with last observed value of this field from previous events example configuration toml inputs fields such a transform could have been implemented by a user with a stateful javascript transform as javascript const fields const storedvalues const handler event for const field of fields if event hasownproperty field storedvalues event else if storedvalues hasownproperty field event storedvalues return event as a bonus this transform would allow to join multiple streams of events for example for this config toml produces events with field cpu temp produces events with field cpu freq type recollect fields inputs fields type console inputs events containing both temperature and frequency would be printed in addition to the most basic form described above we can add an optional configuration parameter forget time which would prevent the field value from being recollected if it was produced earlier than the the current time minus timeout
| 1
|
11,571
| 14,441,672,673
|
IssuesEvent
|
2020-12-07 17:05:34
|
frontendbr/forum
|
https://api.github.com/repos/frontendbr/forum
|
closed
|
Como medir produtividade?
|
Processos [Discussão]
|
Você trabalha home office ou no escritório? Freela ou não? Como medir sua produtividade com tantas variações de modos de trabalho por ai?
Como vocês se avaliam?
|
1.0
|
Como medir produtividade? - Você trabalha home office ou no escritório? Freela ou não? Como medir sua produtividade com tantas variações de modos de trabalho por ai?
Como vocês se avaliam?
|
process
|
como medir produtividade você trabalha home office ou no escritório freela ou não como medir sua produtividade com tantas variações de modos de trabalho por ai como vocês se avaliam
| 1
|
590
| 3,067,143,898
|
IssuesEvent
|
2015-08-18 08:35:47
|
maraujop/django-crispy-forms
|
https://api.github.com/repos/maraujop/django-crispy-forms
|
closed
|
Use py.test for testing
|
Testing/Process
|
Hi!
I want to propose to rewrite test suite with py.test. Main goals are:
* Simplicity. Now test suite is quite complicated and has a lot of code duplication
* Granularity. py.test has beautiful feature to [mark tests](http://pytest.org/latest/parametrize.html#parametrized-test-functions) to run it in different circumstances. Running tests for different templates pack would be easy. Also py.test has [parametrization](http://pytest.org/latest/parametrize.html#parametrized-test-functions), so it should plays great with different templates packs too.
* Split setup and tests. Test code would be more clear and will have less duplication. Also test [fixtures](http://pytest.org/latest/fixture.html#fixture) will be reusable.
It will be better to show the code, so, don't you mind if I'll create some draft for that?
|
1.0
|
Use py.test for testing - Hi!
I want to propose to rewrite test suite with py.test. Main goals are:
* Simplicity. Now test suite is quite complicated and has a lot of code duplication
* Granularity. py.test has beautiful feature to [mark tests](http://pytest.org/latest/parametrize.html#parametrized-test-functions) to run it in different circumstances. Running tests for different templates pack would be easy. Also py.test has [parametrization](http://pytest.org/latest/parametrize.html#parametrized-test-functions), so it should plays great with different templates packs too.
* Split setup and tests. Test code would be more clear and will have less duplication. Also test [fixtures](http://pytest.org/latest/fixture.html#fixture) will be reusable.
It will be better to show the code, so, don't you mind if I'll create some draft for that?
|
process
|
use py test for testing hi i want to propose to rewrite test suite with py test main goals are simplicity now test suite is quite complicated and has a lot of code duplication granularity py test has beautiful feature to to run it in different circumstances running tests for different templates pack would be easy also py test has so it should plays great with different templates packs too split setup and tests test code would be more clear and will have less duplication also test will be reusable it will be better to show the code so don t you mind if i ll create some draft for that
| 1
|
43,318
| 7,042,094,892
|
IssuesEvent
|
2017-12-30 06:30:25
|
zulip/zulip
|
https://api.github.com/repos/zulip/zulip
|
closed
|
user docs: Clarify some items
|
area: documentation (user) enhancement
|
I've noticed a few broken links or inconsistencies in the docs, leading to some possibly unclear parts of the documentation.
- [ ] https://chat.zulip.org/help/getting-started-with-zulip. At the bottom of this page (under `Next Steps`), the **"Add an avatar"** and **"Configure your notifications"** links link directly to the settings, unlike many of the other links in the documentation, which link you to the page that shows how to do those things.
- [ ] https://chat.zulip.org/help/change-your-password. For number 2 under "If you forgot your password", the terminology on the Zulip homepage has been changed to "Forgot your password?" from "Forgot password?"
- [ ] https://chat.zulip.org/help/signing-in. There are more than 3 ways to sign in that aren't documented here.
- [ ] https://chat.zulip.org/help/analytics. This page should be a little bit more clear on how to get to the statistics; it's fairly easy to miss the one (broken) link at the top of the page.
- [ ] https://chat.zulip.org/help/zulip-on-windows. This and the page for Android should be deleted as they seem to be from older versions of those apps. The link for the Windows download is broken on this page, but the Android one links to the correct Google Play Store listing.
- [ ] https://chat.zulip.org/help/about-streams-and-topics. On this page, for the links to customize stream settings, there are many items missing, so something should be added to lead users to the other customization options.
- [ ] https://chat.zulip.org/help/change-your-organization-settings. This one is extremely minor, but an image for a chevron is present next to all the other instances of the word "chevron" in the docs (same with "gear"), but it's missing here.
|
1.0
|
user docs: Clarify some items - I've noticed a few broken links or inconsistencies in the docs, leading to some possibly unclear parts of the documentation.
- [ ] https://chat.zulip.org/help/getting-started-with-zulip. At the bottom of this page (under `Next Steps`), the **"Add an avatar"** and **"Configure your notifications"** links link directly to the settings, unlike many of the other links in the documentation, which link you to the page that shows how to do those things.
- [ ] https://chat.zulip.org/help/change-your-password. For number 2 under "If you forgot your password", the terminology on the Zulip homepage has been changed to "Forgot your password?" from "Forgot password?"
- [ ] https://chat.zulip.org/help/signing-in. There are more than 3 ways to sign in that aren't documented here.
- [ ] https://chat.zulip.org/help/analytics. This page should be a little bit more clear on how to get to the statistics; it's fairly easy to miss the one (broken) link at the top of the page.
- [ ] https://chat.zulip.org/help/zulip-on-windows. This and the page for Android should be deleted as they seem to be from older versions of those apps. The link for the Windows download is broken on this page, but the Android one links to the correct Google Play Store listing.
- [ ] https://chat.zulip.org/help/about-streams-and-topics. On this page, for the links to customize stream settings, there are many items missing, so something should be added to lead users to the other customization options.
- [ ] https://chat.zulip.org/help/change-your-organization-settings. This one is extremely minor, but an image for a chevron is present next to all the other instances of the word "chevron" in the docs (same with "gear"), but it's missing here.
|
non_process
|
user docs clarify some items i ve noticed a few broken links or inconsistencies in the docs leading to some possibly unclear parts of the documentation at the bottom of this page under next steps the add an avatar and configure your notifications links link directly to the settings unlike many of the other links in the documentation which link you to the page that shows how to do those things for number under if you forgot your password the terminology on the zulip homepage has been changed to forgot your password from forgot password there are more than ways to sign in that aren t documented here this page should be a little bit more clear on how to get to the statistics it s fairly easy to miss the one broken link at the top of the page this and the page for android should be deleted as they seem to be from older versions of those apps the link for the windows download is broken on this page but the android one links to the correct google play store listing on this page for the links to customize stream settings there are many items missing so something should be added to lead users to the other customization options this one is extremely minor but an image for a chevron is present next to all the other instances of the word chevron in the docs same with gear but it s missing here
| 0
|
31,036
| 8,641,807,119
|
IssuesEvent
|
2018-11-24 21:54:19
|
Trustroots/trustroots
|
https://api.github.com/repos/Trustroots/trustroots
|
closed
|
easy installation for devs
|
build
|
It seems currently the only way to get this set up is if you have a _lot_ of knowledge of node, docker, and mongo.
It should be possible to set this up in less than 15 minutes with minimal knowledge of any of these, at least on Debian/Ubuntu and OSX.
How can we do this?
|
1.0
|
easy installation for devs - It seems currently the only way to get this set up is if you have a _lot_ of knowledge of node, docker, and mongo.
It should be possible to set this up in less than 15 minutes with minimal knowledge of any of these, at least on Debian/Ubuntu and OSX.
How can we do this?
|
non_process
|
easy installation for devs it seems currently the only way to get this set up is if you have a lot of knowledge of node docker and mongo it should be possible to set this up in less than minutes with minimal knowledge of any of these at least on debian ubuntu and osx how can we do this
| 0
|
299,815
| 22,625,193,509
|
IssuesEvent
|
2022-06-30 10:00:25
|
C3ETH/c3eth-operations
|
https://api.github.com/repos/C3ETH/c3eth-operations
|
closed
|
Catalyst [Challenge Team]: Attending Challenge Team Roundtable 5 meeting
|
Documentation:Catalyst Fund Catalyst:Fund 8 Treasury:Tasks
|
**Tasks:**
- [x] Attend the Roundtable 5 on Friday, 06 May 2022
- [x] Check this [Fund8 Challenge Teams & Onboarding Guidance](https://docs.google.com/document/d/1lmlrqrTthz1CjCj2zNtY7hlAfGsKui_cXIUNygLPsm0/edit)
- [x] Find two available 1.5 hours slot during 18-24 May 2022 for Proof of Life meetings
- [x] Fill this [Google Form](https://docs.google.com/forms/d/e/1FAIpQLSfKrCwNypGi9vT5rCy2eaHKpe-PbKokFp7nftY_OuzhQmxSrg/viewform) for the slot that you want
- [x] Confirmed at least two members of the CT will take part in the Proof of Life meetings
- [x] [Miro board](https://miro.com/app/board/uXjVO4r7A_c=/)
|
1.0
|
Catalyst [Challenge Team]: Attending Challenge Team Roundtable 5 meeting - **Tasks:**
- [x] Attend the Roundtable 5 on Friday, 06 May 2022
- [x] Check this [Fund8 Challenge Teams & Onboarding Guidance](https://docs.google.com/document/d/1lmlrqrTthz1CjCj2zNtY7hlAfGsKui_cXIUNygLPsm0/edit)
- [x] Find two available 1.5 hours slot during 18-24 May 2022 for Proof of Life meetings
- [x] Fill this [Google Form](https://docs.google.com/forms/d/e/1FAIpQLSfKrCwNypGi9vT5rCy2eaHKpe-PbKokFp7nftY_OuzhQmxSrg/viewform) for the slot that you want
- [x] Confirmed at least two members of the CT will take part in the Proof of Life meetings
- [x] [Miro board](https://miro.com/app/board/uXjVO4r7A_c=/)
|
non_process
|
catalyst attending challenge team roundtable meeting tasks attend the roundtable on friday may check this find two available hours slot during may for proof of life meetings fill this for the slot that you want confirmed at least two members of the ct will take part in the proof of life meetings
| 0
|
308,613
| 23,257,415,216
|
IssuesEvent
|
2022-08-04 10:29:53
|
neogulcity/jsonparser
|
https://api.github.com/repos/neogulcity/jsonparser
|
opened
|
데이터 타입 판단
|
documentation
|
```c++
int myInt[2];
int* myInt = new int[2];
```
int[], int* 두 가지 데이터 타입에 따른 Json 배열 처리 방법이 필요함.
|
1.0
|
데이터 타입 판단 - ```c++
int myInt[2];
int* myInt = new int[2];
```
int[], int* 두 가지 데이터 타입에 따른 Json 배열 처리 방법이 필요함.
|
non_process
|
데이터 타입 판단 c int myint int myint new int int int 두 가지 데이터 타입에 따른 json 배열 처리 방법이 필요함
| 0
|
4,550
| 7,375,395,204
|
IssuesEvent
|
2018-03-14 00:13:09
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
Consistency in <acrName>
|
assigned-to-author container-service doc-enhancement in-process triaged
|
In the previous steps (1 and 2), you use <acrName> as a variable. In the below it's "myACRRegistry". Suggest to replace by <acrName>
ACR_ID=$(az acr show --name myACRRegistry --resource-group myResourceGroup --query "id" --output tsv)
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: d4af71a4-648e-580f-7e86-94ce05196de5
* Version Independent ID: 134b8ba0-b085-076a-d2e1-16d0ddff5586
* Content: [Kubernetes on Azure tutorial - Deploy Cluster | Microsoft Docs](https://docs.microsoft.com/en-us/azure/aks/tutorial-kubernetes-deploy-cluster)
* Content Source: [articles/aks/tutorial-kubernetes-deploy-cluster.md](https://github.com/Microsoft/azure-docs/blob/master/articles/aks/tutorial-kubernetes-deploy-cluster.md)
* Service: **container-service**
* GitHub Login: @neilpeterson
* Microsoft Alias: **nepeters**
|
1.0
|
Consistency in <acrName> - In the previous steps (1 and 2), you use <acrName> as a variable. In the below it's "myACRRegistry". Suggest to replace by <acrName>
ACR_ID=$(az acr show --name myACRRegistry --resource-group myResourceGroup --query "id" --output tsv)
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: d4af71a4-648e-580f-7e86-94ce05196de5
* Version Independent ID: 134b8ba0-b085-076a-d2e1-16d0ddff5586
* Content: [Kubernetes on Azure tutorial - Deploy Cluster | Microsoft Docs](https://docs.microsoft.com/en-us/azure/aks/tutorial-kubernetes-deploy-cluster)
* Content Source: [articles/aks/tutorial-kubernetes-deploy-cluster.md](https://github.com/Microsoft/azure-docs/blob/master/articles/aks/tutorial-kubernetes-deploy-cluster.md)
* Service: **container-service**
* GitHub Login: @neilpeterson
* Microsoft Alias: **nepeters**
|
process
|
consistency in in the previous steps and you use as a variable in the below it s myacrregistry suggest to replace by acr id az acr show name myacrregistry resource group myresourcegroup query id output tsv document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service container service github login neilpeterson microsoft alias nepeters
| 1
|
19,460
| 25,752,788,667
|
IssuesEvent
|
2022-12-08 14:23:56
|
influxdata/telegraf
|
https://api.github.com/repos/influxdata/telegraf
|
closed
|
processors.parser changes measurement name to 'parser'
|
bug plugin/parser plugin/processor
|
### Relevant telegraf.conf
```toml
[[inputs.file]]
files = ["/test/test.txt"]
data_format = "value"
data_type = "string"
[[processors.parser]]
namepass = ["file"]
parse_fields = ["value"]
drop_original = false
merge = "override"
data_format = "value"
data_type = "float"
[[outputs.file]]
files = ["stdout"]
```
### Logs from Telegraf
```text
Telegraf 1.23.4:
file value=7.2 1664873970000000000
Telegraf 1.24.2:
parser value=7.2 1664892267000000000
```
### Steps to reproduce
Use the [[processors.parser]] as shown in the example above or as reported in one of links given in the "Additional info" block.
### Expected behavior
like Telegraf 1.23.4 does: keep the measurement name of the input.
### Actual behavior
Measurement name is changed to 'parser'.
The field is dropped from the original measurement, even with setting `drop_original = false`.
### Additional info
The same issue get's talked about behind these 3 links:
* https://community.influxdata.com/t/parser-processor-change-measurement-name-to-parser/26984
* https://community.influxdata.com/t/single-mqtt-consumer-for-int-float-string/27137
* https://influxcommunity.slack.com/archives/CH99HUH8V/p1666794045416509
I'm a new user of InfluxDB & Telegraf, and have no idea how to actually switch to an older Telegraf version. I've copied the sample and it's output from this issue:
https://github.com/influxdata/telegraf/issues/11921
|
1.0
|
processors.parser changes measurement name to 'parser' - ### Relevant telegraf.conf
```toml
[[inputs.file]]
files = ["/test/test.txt"]
data_format = "value"
data_type = "string"
[[processors.parser]]
namepass = ["file"]
parse_fields = ["value"]
drop_original = false
merge = "override"
data_format = "value"
data_type = "float"
[[outputs.file]]
files = ["stdout"]
```
### Logs from Telegraf
```text
Telegraf 1.23.4:
file value=7.2 1664873970000000000
Telegraf 1.24.2:
parser value=7.2 1664892267000000000
```
### Steps to reproduce
Use the [[processors.parser]] as shown in the example above or as reported in one of links given in the "Additional info" block.
### Expected behavior
like Telegraf 1.23.4 does: keep the measurement name of the input.
### Actual behavior
Measurement name is changed to 'parser'.
The field is dropped from the original measurement, even with setting `drop_original = false`.
### Additional info
The same issue get's talked about behind these 3 links:
* https://community.influxdata.com/t/parser-processor-change-measurement-name-to-parser/26984
* https://community.influxdata.com/t/single-mqtt-consumer-for-int-float-string/27137
* https://influxcommunity.slack.com/archives/CH99HUH8V/p1666794045416509
I'm a new user of InfluxDB & Telegraf, and have no idea how to actually switch to an older Telegraf version. I've copied the sample and it's output from this issue:
https://github.com/influxdata/telegraf/issues/11921
|
process
|
processors parser changes measurement name to parser relevant telegraf conf toml files data format value data type string namepass parse fields drop original false merge override data format value data type float files logs from telegraf text telegraf file value telegraf parser value steps to reproduce use the as shown in the example above or as reported in one of links given in the additional info block expected behavior like telegraf does keep the measurement name of the input actual behavior measurement name is changed to parser the field is dropped from the original measurement even with setting drop original false additional info the same issue get s talked about behind these links i m a new user of influxdb telegraf and have no idea how to actually switch to an older telegraf version i ve copied the sample and it s output from this issue
| 1
|
260,934
| 8,221,308,300
|
IssuesEvent
|
2018-09-06 01:13:06
|
C3DSU/e-DefPR
|
https://api.github.com/repos/C3DSU/e-DefPR
|
closed
|
Padronizar nomes de atributos do backend
|
Category: Backend Priority: Medium Type: Bug
|
Padronizar nomes de atributos do backend, não utilizar mais CamelCase
|
1.0
|
Padronizar nomes de atributos do backend - Padronizar nomes de atributos do backend, não utilizar mais CamelCase
|
non_process
|
padronizar nomes de atributos do backend padronizar nomes de atributos do backend não utilizar mais camelcase
| 0
|
17,746
| 23,659,317,936
|
IssuesEvent
|
2022-08-26 14:10:30
|
saibrotech/mentoria
|
https://api.github.com/repos/saibrotech/mentoria
|
closed
|
Fazer processo seletivo JaBrasil TI
|
processo seletivo
|
https://jabrasil.org.br/ti
Santa Catarina
Aulas: Aulas online e ao vivo. Exercícios e conteúdos na plataforma Coursera. Você precisará se dedicar 5 vezes por semana, 4 horas por dia. Início das aulas: 29.08 a 09.12
Número de Vagas: 20
- [x] Fazer inscrição
- [x] Enviar documentação necessária
- [x] Participar dos encontros via Meet nos dias 22, 23 e 24 de agosto às 19h
- [x] Compartilhar certificado da semana experimental no LinkedIn
- [x] Aguardar resultado
|
1.0
|
Fazer processo seletivo JaBrasil TI - https://jabrasil.org.br/ti
Santa Catarina
Aulas: Aulas online e ao vivo. Exercícios e conteúdos na plataforma Coursera. Você precisará se dedicar 5 vezes por semana, 4 horas por dia. Início das aulas: 29.08 a 09.12
Número de Vagas: 20
- [x] Fazer inscrição
- [x] Enviar documentação necessária
- [x] Participar dos encontros via Meet nos dias 22, 23 e 24 de agosto às 19h
- [x] Compartilhar certificado da semana experimental no LinkedIn
- [x] Aguardar resultado
|
process
|
fazer processo seletivo jabrasil ti santa catarina aulas aulas online e ao vivo exercícios e conteúdos na plataforma coursera você precisará se dedicar vezes por semana horas por dia início das aulas a número de vagas fazer inscrição enviar documentação necessária participar dos encontros via meet nos dias e de agosto às compartilhar certificado da semana experimental no linkedin aguardar resultado
| 1
|
37,011
| 9,941,105,857
|
IssuesEvent
|
2019-07-03 10:44:05
|
ShaikASK/Testing
|
https://api.github.com/repos/ShaikASK/Testing
|
closed
|
Design Webform : Custom Webform : User is able to Create more then one custom webforms with same name
|
Custom Webforms Defect Design Webform HR Admin Module P2 Release #4 Build # 3
|
Steps To Replicate :
1.Launch the URL
2.Sign in “HR Admin” user
3.Go to Design Webforms
4.Create a “Custom Webform “ with some name say “Test” & save it
5. Create another webform with same name “Test” and save it
Experienced Behavior : Observed that user is able to Create more then one custom web forms with same name
(Refer Screen Shot)
Expected Behavior : Ensure that application should not allow to create custom web forms with same name

|
1.0
|
Design Webform : Custom Webform : User is able to Create more then one custom webforms with same name - Steps To Replicate :
1.Launch the URL
2.Sign in “HR Admin” user
3.Go to Design Webforms
4.Create a “Custom Webform “ with some name say “Test” & save it
5. Create another webform with same name “Test” and save it
Experienced Behavior : Observed that user is able to Create more then one custom web forms with same name
(Refer Screen Shot)
Expected Behavior : Ensure that application should not allow to create custom web forms with same name

|
non_process
|
design webform custom webform user is able to create more then one custom webforms with same name steps to replicate launch the url sign in “hr admin” user go to design webforms create a “custom webform “ with some name say “test” save it create another webform with same name “test” and save it experienced behavior observed that user is able to create more then one custom web forms with same name refer screen shot expected behavior ensure that application should not allow to create custom web forms with same name
| 0
|
728,572
| 25,084,885,234
|
IssuesEvent
|
2022-11-07 22:44:13
|
yugabyte/yugabyte-db
|
https://api.github.com/repos/yugabyte/yugabyte-db
|
closed
|
[YSQL] Add yb_pg_select_distinct.sql based on select_distinct.sql
|
kind/bug area/ysql priority/medium status/awaiting-triage
|
Jira Link: [DB-4062](https://yugabyte.atlassian.net/browse/DB-4062)
### Description
This issue adds yb_select_distinct.sql which is based on select_distinct.sql
The newly added file would be used for future backports involving select_distinct.sql
|
1.0
|
[YSQL] Add yb_pg_select_distinct.sql based on select_distinct.sql - Jira Link: [DB-4062](https://yugabyte.atlassian.net/browse/DB-4062)
### Description
This issue adds yb_select_distinct.sql which is based on select_distinct.sql
The newly added file would be used for future backports involving select_distinct.sql
|
non_process
|
add yb pg select distinct sql based on select distinct sql jira link description this issue adds yb select distinct sql which is based on select distinct sql the newly added file would be used for future backports involving select distinct sql
| 0
|
446,701
| 12,876,982,403
|
IssuesEvent
|
2020-07-11 08:20:39
|
threefoldtech/js-sdk
|
https://api.github.com/repos/threefoldtech/js-sdk
|
opened
|
starting nginx through starting threebot fails on osx
|
priority_major type_bug
|
psutil problem:
ttps://github.com/giampaolo/psutil/issues/1447#issuecomment-646294879
```
File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/bin/threebot", line 11, in <module>
load_entry_point('js-sdk', 'console_scripts', 'threebot')()
File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/lib/python3.8/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/lib/python3.8/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/lib/python3.8/site-packages/click/core.py", line 1259, in invoke
return _process_result(sub_ctx.command.invoke(sub_ctx))
File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/lib/python3.8/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/lib/python3.8/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/Users/rob/workspace/github.com/threefoldtech/js-sdk/jumpscale/entry_points/threebot.py", line 124, in start
j.servers.threebot.start_default(wait=True, local=local)
File "/Users/rob/workspace/github.com/threefoldtech/js-sdk/jumpscale/servers/threebot/__init__.py", line 29, in start_default
server.start(wait=wait)
File "/Users/rob/workspace/github.com/threefoldtech/js-sdk/jumpscale/servers/threebot/threebot.py", line 593, in start
if self.is_running():
File "/Users/rob/workspace/github.com/threefoldtech/js-sdk/jumpscale/servers/threebot/threebot.py", line 506, in is_running
nginx_running = self.nginx.is_running()
File "/Users/rob/workspace/github.com/threefoldtech/js-sdk/jumpscale/tools/nginx/nginxserver.py", line 54, in is_running
return j.tools.startupcmd.get(f"nginx_{self.name}").is_running()
File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/src/js-ng/jumpscale/tools/startupcmd/startupcmd.py", line 225, in is_running
if self.process:
File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/src/js-ng/jumpscale/tools/startupcmd/startupcmd.py", line 89, in process
if self.pid:
File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/src/js-ng/jumpscale/tools/startupcmd/startupcmd.py", line 70, in pid
pids = j.sals.process.get_pids(f"startupcmd_{self.instance_name}")
File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/src/js-ng/jumpscale/sals/process/__init__.py", line 379, in get_pids
elif match_predicate(process_name, process.exe()):
File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/lib/python3.8/site-packages/psutil/__init__.py", line 673, in exe
exe = self._proc.exe()
File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/lib/python3.8/site-packages/psutil/_psosx.py", line 334, in wrapper
return fun(self, *args, **kwargs)
File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/lib/python3.8/site-packages/psutil/_psosx.py", line 413, in exe
return cext.proc_exe(self.pid)
FileNotFoundError: [Errno 2] No such file or directory (originated from proc_pidpath())
```
|
1.0
|
starting nginx through starting threebot fails on osx -
psutil problem:
ttps://github.com/giampaolo/psutil/issues/1447#issuecomment-646294879
```
File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/bin/threebot", line 11, in <module>
load_entry_point('js-sdk', 'console_scripts', 'threebot')()
File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/lib/python3.8/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/lib/python3.8/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/lib/python3.8/site-packages/click/core.py", line 1259, in invoke
return _process_result(sub_ctx.command.invoke(sub_ctx))
File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/lib/python3.8/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/lib/python3.8/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/Users/rob/workspace/github.com/threefoldtech/js-sdk/jumpscale/entry_points/threebot.py", line 124, in start
j.servers.threebot.start_default(wait=True, local=local)
File "/Users/rob/workspace/github.com/threefoldtech/js-sdk/jumpscale/servers/threebot/__init__.py", line 29, in start_default
server.start(wait=wait)
File "/Users/rob/workspace/github.com/threefoldtech/js-sdk/jumpscale/servers/threebot/threebot.py", line 593, in start
if self.is_running():
File "/Users/rob/workspace/github.com/threefoldtech/js-sdk/jumpscale/servers/threebot/threebot.py", line 506, in is_running
nginx_running = self.nginx.is_running()
File "/Users/rob/workspace/github.com/threefoldtech/js-sdk/jumpscale/tools/nginx/nginxserver.py", line 54, in is_running
return j.tools.startupcmd.get(f"nginx_{self.name}").is_running()
File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/src/js-ng/jumpscale/tools/startupcmd/startupcmd.py", line 225, in is_running
if self.process:
File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/src/js-ng/jumpscale/tools/startupcmd/startupcmd.py", line 89, in process
if self.pid:
File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/src/js-ng/jumpscale/tools/startupcmd/startupcmd.py", line 70, in pid
pids = j.sals.process.get_pids(f"startupcmd_{self.instance_name}")
File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/src/js-ng/jumpscale/sals/process/__init__.py", line 379, in get_pids
elif match_predicate(process_name, process.exe()):
File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/lib/python3.8/site-packages/psutil/__init__.py", line 673, in exe
exe = self._proc.exe()
File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/lib/python3.8/site-packages/psutil/_psosx.py", line 334, in wrapper
return fun(self, *args, **kwargs)
File "/Users/rob/Library/Caches/pypoetry/virtualenvs/js-sdk-Dbu3Jec3-py3.8/lib/python3.8/site-packages/psutil/_psosx.py", line 413, in exe
return cext.proc_exe(self.pid)
FileNotFoundError: [Errno 2] No such file or directory (originated from proc_pidpath())
```
|
non_process
|
starting nginx through starting threebot fails on osx psutil problem ttps github com giampaolo psutil issues issuecomment file users rob library caches pypoetry virtualenvs js sdk bin threebot line in load entry point js sdk console scripts threebot file users rob library caches pypoetry virtualenvs js sdk lib site packages click core py line in call return self main args kwargs file users rob library caches pypoetry virtualenvs js sdk lib site packages click core py line in main rv self invoke ctx file users rob library caches pypoetry virtualenvs js sdk lib site packages click core py line in invoke return process result sub ctx command invoke sub ctx file users rob library caches pypoetry virtualenvs js sdk lib site packages click core py line in invoke return ctx invoke self callback ctx params file users rob library caches pypoetry virtualenvs js sdk lib site packages click core py line in invoke return callback args kwargs file users rob workspace github com threefoldtech js sdk jumpscale entry points threebot py line in start j servers threebot start default wait true local local file users rob workspace github com threefoldtech js sdk jumpscale servers threebot init py line in start default server start wait wait file users rob workspace github com threefoldtech js sdk jumpscale servers threebot threebot py line in start if self is running file users rob workspace github com threefoldtech js sdk jumpscale servers threebot threebot py line in is running nginx running self nginx is running file users rob workspace github com threefoldtech js sdk jumpscale tools nginx nginxserver py line in is running return j tools startupcmd get f nginx self name is running file users rob library caches pypoetry virtualenvs js sdk src js ng jumpscale tools startupcmd startupcmd py line in is running if self process file users rob library caches pypoetry virtualenvs js sdk src js ng jumpscale tools startupcmd startupcmd py line in process if self pid file users rob library caches pypoetry virtualenvs js sdk src js ng jumpscale tools startupcmd startupcmd py line in pid pids j sals process get pids f startupcmd self instance name file users rob library caches pypoetry virtualenvs js sdk src js ng jumpscale sals process init py line in get pids elif match predicate process name process exe file users rob library caches pypoetry virtualenvs js sdk lib site packages psutil init py line in exe exe self proc exe file users rob library caches pypoetry virtualenvs js sdk lib site packages psutil psosx py line in wrapper return fun self args kwargs file users rob library caches pypoetry virtualenvs js sdk lib site packages psutil psosx py line in exe return cext proc exe self pid filenotfounderror no such file or directory originated from proc pidpath
| 0
|
66,023
| 12,703,479,346
|
IssuesEvent
|
2020-06-22 22:25:33
|
dotnet/runtime
|
https://api.github.com/repos/dotnet/runtime
|
opened
|
Attribute and analyzer for "soft abstract"
|
api-ready-for-review code-analyzer code-fixer
|
For abstract base types, we sometimes find ourselves in a situation where we'd like to add a new core concept to a type. In order to stay ABI-compatible, we can't add new `abstract` members as this means derived types that were compiled against the earlier version are missing the override, thus causing a `TypeLoadException`.
The general pattern for this that we make the new members `virtual`, ideally with a workable implementation. This doesn't always work, so sometimes the only option is adding an implementation that throws `NotImplementedException`. A good example of this problem is in https://github.com/dotnet/runtime/issues/31362#issuecomment-643438552.
A major downside of this approach is that developers deriving from the base type don't necessarily know that they are supposed to override these new virtual methods to get the best behavior.
One option to solve this problem is by adding a new attribute and analyzer that warns when deriving from it without overriding the marked members.
### API Proposal
```C#
namespace System.Diagnostics.CodeAnalysis
{
[AttributeUsage(AttributeTargets.Event |
AttributeTargets.Method |
AttributeTargets.Property)]
public sealed class ShouldBeOverridenAttribute : Attribute
{
public ShouldBeOverridenAttribute();
}
}
```
### Analyzer Proposal
Let's take [this example from the BCL](https://github.com/dotnet/runtime/issues/31362#issuecomment-643438552) and assume we marked the new virtual members as `[ShouldBeOverriden]` like this:
```C#
namespace System.Threading.Channels
{
public abstract partial class ChannelReader<T>
{
[ShouldBeOverriden]
public virtual bool CanPeek => false;
[ShouldBeOverriden]
public virtual bool TryPeek([MaybeNullWhen(false)] out T item)
{
item = default!;
return false;
}
}
}
```
A customer deriving from `ChannelReader<T>` would only override the `abstract` members by default, resulting in this:
```C#
public class MyChannelReader<T> : ChannelReader<T>
~~~~~~~~~~~~~~~
{
public override bool TryRead([MaybeNullWhen(false)] out T item)
{
throw new NotImplementedException();
}
public override ValueTask<bool> WaitToReadAsync(CancellationToken cancellationToken = default)
{
throw new NotImplementedException();
}
}
```
This would product warnings on the type delaration:
```text
WARNING: ChannelReader<T>.CanPeek should be overriden
WARNING: ChannelReader<T>.TryPeek(out T item) should be overriden
```
### IDE
We may want to change the behavior of the IDE so that the code fixer that generates overrides for abstract members automatically also generates overrides for any members marked as `[ShouldBeOverriden]`.
/cc @dotnet/fxdc @CyrusNajmabadi
|
2.0
|
Attribute and analyzer for "soft abstract" - For abstract base types, we sometimes find ourselves in a situation where we'd like to add a new core concept to a type. In order to stay ABI-compatible, we can't add new `abstract` members as this means derived types that were compiled against the earlier version are missing the override, thus causing a `TypeLoadException`.
The general pattern for this that we make the new members `virtual`, ideally with a workable implementation. This doesn't always work, so sometimes the only option is adding an implementation that throws `NotImplementedException`. A good example of this problem is in https://github.com/dotnet/runtime/issues/31362#issuecomment-643438552.
A major downside of this approach is that developers deriving from the base type don't necessarily know that they are supposed to override these new virtual methods to get the best behavior.
One option to solve this problem is by adding a new attribute and analyzer that warns when deriving from it without overriding the marked members.
### API Proposal
```C#
namespace System.Diagnostics.CodeAnalysis
{
[AttributeUsage(AttributeTargets.Event |
AttributeTargets.Method |
AttributeTargets.Property)]
public sealed class ShouldBeOverridenAttribute : Attribute
{
public ShouldBeOverridenAttribute();
}
}
```
### Analyzer Proposal
Let's take [this example from the BCL](https://github.com/dotnet/runtime/issues/31362#issuecomment-643438552) and assume we marked the new virtual members as `[ShouldBeOverriden]` like this:
```C#
namespace System.Threading.Channels
{
public abstract partial class ChannelReader<T>
{
[ShouldBeOverriden]
public virtual bool CanPeek => false;
[ShouldBeOverriden]
public virtual bool TryPeek([MaybeNullWhen(false)] out T item)
{
item = default!;
return false;
}
}
}
```
A customer deriving from `ChannelReader<T>` would only override the `abstract` members by default, resulting in this:
```C#
public class MyChannelReader<T> : ChannelReader<T>
~~~~~~~~~~~~~~~
{
public override bool TryRead([MaybeNullWhen(false)] out T item)
{
throw new NotImplementedException();
}
public override ValueTask<bool> WaitToReadAsync(CancellationToken cancellationToken = default)
{
throw new NotImplementedException();
}
}
```
This would product warnings on the type delaration:
```text
WARNING: ChannelReader<T>.CanPeek should be overriden
WARNING: ChannelReader<T>.TryPeek(out T item) should be overriden
```
### IDE
We may want to change the behavior of the IDE so that the code fixer that generates overrides for abstract members automatically also generates overrides for any members marked as `[ShouldBeOverriden]`.
/cc @dotnet/fxdc @CyrusNajmabadi
|
non_process
|
attribute and analyzer for soft abstract for abstract base types we sometimes find ourselves in a situation where we d like to add a new core concept to a type in order to stay abi compatible we can t add new abstract members as this means derived types that were compiled against the earlier version are missing the override thus causing a typeloadexception the general pattern for this that we make the new members virtual ideally with a workable implementation this doesn t always work so sometimes the only option is adding an implementation that throws notimplementedexception a good example of this problem is in a major downside of this approach is that developers deriving from the base type don t necessarily know that they are supposed to override these new virtual methods to get the best behavior one option to solve this problem is by adding a new attribute and analyzer that warns when deriving from it without overriding the marked members api proposal c namespace system diagnostics codeanalysis attributeusage attributetargets event attributetargets method attributetargets property public sealed class shouldbeoverridenattribute attribute public shouldbeoverridenattribute analyzer proposal let s take and assume we marked the new virtual members as like this c namespace system threading channels public abstract partial class channelreader public virtual bool canpeek false public virtual bool trypeek out t item item default return false a customer deriving from channelreader would only override the abstract members by default resulting in this c public class mychannelreader channelreader public override bool tryread out t item throw new notimplementedexception public override valuetask waittoreadasync cancellationtoken cancellationtoken default throw new notimplementedexception this would product warnings on the type delaration text warning channelreader canpeek should be overriden warning channelreader trypeek out t item should be overriden ide we may want to change the behavior of the ide so that the code fixer that generates overrides for abstract members automatically also generates overrides for any members marked as cc dotnet fxdc cyrusnajmabadi
| 0
|
442,803
| 12,750,625,150
|
IssuesEvent
|
2020-06-27 05:50:44
|
VandyHacks/vaken
|
https://api.github.com/repos/VandyHacks/vaken
|
opened
|
Add Summer Hackathon Waiver to the VH Summer 2020 Vaken Portal
|
Priority: High bite size good first issue
|
Add the summer hackathon waiver to application in the `event/vh2020-summer` branch.
|
1.0
|
Add Summer Hackathon Waiver to the VH Summer 2020 Vaken Portal - Add the summer hackathon waiver to application in the `event/vh2020-summer` branch.
|
non_process
|
add summer hackathon waiver to the vh summer vaken portal add the summer hackathon waiver to application in the event summer branch
| 0
|
127,762
| 17,360,351,711
|
IssuesEvent
|
2021-07-29 19:39:42
|
CMSgov/eAPD
|
https://api.github.com/repos/CMSgov/eAPD
|
opened
|
[Design Issue] Design a pattern for the export view for when non-required fields are not filled out
|
design
|
Some of the fields in the eAPD will not be required to be answered before submit. We have language that we use, noted in the content guide in the wiki, to note when a field isn't filled out so that it's clear to the reviewer that it wasn't a mistake or something broken in the system. Determine if we need any more language patterns or visual patterns to show when fields are not answered.
### This task is done when…
- [ ] any acceptance criteria (not process oriented, requirements of feature)
- [ ] designs are created, taking into consideration multiple FFY, multiple entries, long entries, weird entries, and/or the export view if applicable, validation considereations
- [ ] 508 considerations and/or recommendations are reviewed
- [ ] designs are shared
- [ ] designs are approved by product
- [ ] if there is a long discussion, a summary of changes for dev will be added to the last comment before transfer
- [ ] a list of changes to be made if this is improving an existing feature
---
Add additional labels (design, dev, compliance, BUG, etc) and size before submitting.
If the issue is needed to complete prioritized work for the CURRENT SPRINT, add it to the "This Sprint" pipeline. Otherwise, all other issues will be automatically added to the _unprioritized pipeline_ for prioritization in backlog refinement or sprint planning with Product.
|
1.0
|
[Design Issue] Design a pattern for the export view for when non-required fields are not filled out - Some of the fields in the eAPD will not be required to be answered before submit. We have language that we use, noted in the content guide in the wiki, to note when a field isn't filled out so that it's clear to the reviewer that it wasn't a mistake or something broken in the system. Determine if we need any more language patterns or visual patterns to show when fields are not answered.
### This task is done when…
- [ ] any acceptance criteria (not process oriented, requirements of feature)
- [ ] designs are created, taking into consideration multiple FFY, multiple entries, long entries, weird entries, and/or the export view if applicable, validation considereations
- [ ] 508 considerations and/or recommendations are reviewed
- [ ] designs are shared
- [ ] designs are approved by product
- [ ] if there is a long discussion, a summary of changes for dev will be added to the last comment before transfer
- [ ] a list of changes to be made if this is improving an existing feature
---
Add additional labels (design, dev, compliance, BUG, etc) and size before submitting.
If the issue is needed to complete prioritized work for the CURRENT SPRINT, add it to the "This Sprint" pipeline. Otherwise, all other issues will be automatically added to the _unprioritized pipeline_ for prioritization in backlog refinement or sprint planning with Product.
|
non_process
|
design a pattern for the export view for when non required fields are not filled out some of the fields in the eapd will not be required to be answered before submit we have language that we use noted in the content guide in the wiki to note when a field isn t filled out so that it s clear to the reviewer that it wasn t a mistake or something broken in the system determine if we need any more language patterns or visual patterns to show when fields are not answered this task is done when… any acceptance criteria not process oriented requirements of feature designs are created taking into consideration multiple ffy multiple entries long entries weird entries and or the export view if applicable validation considereations considerations and or recommendations are reviewed designs are shared designs are approved by product if there is a long discussion a summary of changes for dev will be added to the last comment before transfer a list of changes to be made if this is improving an existing feature add additional labels design dev compliance bug etc and size before submitting if the issue is needed to complete prioritized work for the current sprint add it to the this sprint pipeline otherwise all other issues will be automatically added to the unprioritized pipeline for prioritization in backlog refinement or sprint planning with product
| 0
|
45,749
| 5,730,838,307
|
IssuesEvent
|
2017-04-21 10:31:01
|
missioncommand/emp3-android
|
https://api.github.com/repos/missioncommand/emp3-android
|
closed
|
Test Tactial Graphics
|
integration testing Test Capability
|
Test drawing editing and renderering of the following MilStd tactical graphics in development.
- [ ] Bridge => WAR.GRDTRK.EQT.GRDVEH.ENGVEH.BRG
- [ ] bridge => TACTICAL GRAPHICS / MOBILITY/SURVIVABILITY / OBSTACLE BYPASS / CROSSING SITE/WATER CROSSING / BRIDGE OR GAP
- [ ] Attack by Fire => TACTICAL GRAPHICS / COMMAND AND CONTROL AND GENERAL MANEUVER / OFFENSE / AREAS / ATTACK BY FIRE POSITION
- [ ] Support by fire position
=> TACTICAL GRAPHICS / COMMAND AND CONTROL AND GENERAL MANEUVER / OFFENSE / AREAS / SUPPORT BY FIRE POSITION
|
2.0
|
Test Tactial Graphics - Test drawing editing and renderering of the following MilStd tactical graphics in development.
- [ ] Bridge => WAR.GRDTRK.EQT.GRDVEH.ENGVEH.BRG
- [ ] bridge => TACTICAL GRAPHICS / MOBILITY/SURVIVABILITY / OBSTACLE BYPASS / CROSSING SITE/WATER CROSSING / BRIDGE OR GAP
- [ ] Attack by Fire => TACTICAL GRAPHICS / COMMAND AND CONTROL AND GENERAL MANEUVER / OFFENSE / AREAS / ATTACK BY FIRE POSITION
- [ ] Support by fire position
=> TACTICAL GRAPHICS / COMMAND AND CONTROL AND GENERAL MANEUVER / OFFENSE / AREAS / SUPPORT BY FIRE POSITION
|
non_process
|
test tactial graphics test drawing editing and renderering of the following milstd tactical graphics in development bridge war grdtrk eqt grdveh engveh brg bridge tactical graphics mobility survivability obstacle bypass crossing site water crossing bridge or gap attack by fire tactical graphics command and control and general maneuver offense areas attack by fire position support by fire position tactical graphics command and control and general maneuver offense areas support by fire position
| 0
|
10,258
| 13,110,947,551
|
IssuesEvent
|
2020-08-04 21:45:16
|
SCIInstitute/Seg3D
|
https://api.github.com/repos/SCIInstitute/Seg3D
|
opened
|
Investigate VTK to replace OpenGL
|
graphics renderer software processes
|
Seg3D viewer features:
Opacity
Slice viewing
Isosurfacing
Brightness - gamma functions
Colormaps
Real-time feedback for labels
Texture mapping
Heads-up display
Interactive feedback for paint, polylines, and other seed selection
Volume rendering
Optional: 3D interaction - seed point picking, moving, painting
|
1.0
|
Investigate VTK to replace OpenGL - Seg3D viewer features:
Opacity
Slice viewing
Isosurfacing
Brightness - gamma functions
Colormaps
Real-time feedback for labels
Texture mapping
Heads-up display
Interactive feedback for paint, polylines, and other seed selection
Volume rendering
Optional: 3D interaction - seed point picking, moving, painting
|
process
|
investigate vtk to replace opengl viewer features opacity slice viewing isosurfacing brightness gamma functions colormaps real time feedback for labels texture mapping heads up display interactive feedback for paint polylines and other seed selection volume rendering optional interaction seed point picking moving painting
| 1
|
168,856
| 6,388,170,740
|
IssuesEvent
|
2017-08-03 15:04:20
|
fossasia/susi_skill_cms
|
https://api.github.com/repos/fossasia/susi_skill_cms
|
closed
|
Fix Dropdown menu structure and views across apps and pages
|
bug Priority: High
|
Related to: https://github.com/fossasia/chat.susi.ai/issues/657
**Not logged in users:**
About -> Overview
Chat -> /chat app
Skills -> / skills app
Settings -> /settings
Login -> login
**Logged in users:**
About -> Overview
Chat -> /chat app
Skills -> / skills app
Settings -> /settings
[Themes -> theme popup (only in chat app)]
Logout -> logout
* Add icons to all menu items
* Don't show "signup" on menu. The link to signup on the login page is sufficient.
* On chat app direct settings to chat app settings
* On skills cms direct settings link to skills settings
* Ensure the dropdown menu works on the same way on chat pages overview etc.
|
1.0
|
Fix Dropdown menu structure and views across apps and pages - Related to: https://github.com/fossasia/chat.susi.ai/issues/657
**Not logged in users:**
About -> Overview
Chat -> /chat app
Skills -> / skills app
Settings -> /settings
Login -> login
**Logged in users:**
About -> Overview
Chat -> /chat app
Skills -> / skills app
Settings -> /settings
[Themes -> theme popup (only in chat app)]
Logout -> logout
* Add icons to all menu items
* Don't show "signup" on menu. The link to signup on the login page is sufficient.
* On chat app direct settings to chat app settings
* On skills cms direct settings link to skills settings
* Ensure the dropdown menu works on the same way on chat pages overview etc.
|
non_process
|
fix dropdown menu structure and views across apps and pages related to not logged in users about overview chat chat app skills skills app settings settings login login logged in users about overview chat chat app skills skills app settings settings logout logout add icons to all menu items don t show signup on menu the link to signup on the login page is sufficient on chat app direct settings to chat app settings on skills cms direct settings link to skills settings ensure the dropdown menu works on the same way on chat pages overview etc
| 0
|
192,495
| 6,850,544,086
|
IssuesEvent
|
2017-11-14 03:55:54
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
hacks.mozilla.org - see bug description
|
browser-firefox priority-important
|
<!-- @browser: Firefox 57.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:57.0) Gecko/20100101 Firefox/57.0 -->
<!-- @reported_with: web -->
**URL**: https://hacks.mozilla.org/
**Browser / Version**: Firefox 57.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes
**Problem type**: Something else
**Description**: When I type the letter Y in the search bar it looks like a V.
**Steps to Reproduce**:
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
hacks.mozilla.org - see bug description - <!-- @browser: Firefox 57.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:57.0) Gecko/20100101 Firefox/57.0 -->
<!-- @reported_with: web -->
**URL**: https://hacks.mozilla.org/
**Browser / Version**: Firefox 57.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes
**Problem type**: Something else
**Description**: When I type the letter Y in the search bar it looks like a V.
**Steps to Reproduce**:
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
hacks mozilla org see bug description url browser version firefox operating system windows tested another browser yes problem type something else description when i type the letter y in the search bar it looks like a v steps to reproduce from with ❤️
| 0
|
20,056
| 26,542,945,083
|
IssuesEvent
|
2023-01-19 20:56:13
|
CSE201-project/PaperFriend-desktop-app
|
https://api.github.com/repos/CSE201-project/PaperFriend-desktop-app
|
closed
|
write an adapting welcome msg everytime the user log in
|
enhancement file processing
|
this msg will be different depending on how long the user was log out
|
1.0
|
write an adapting welcome msg everytime the user log in - this msg will be different depending on how long the user was log out
|
process
|
write an adapting welcome msg everytime the user log in this msg will be different depending on how long the user was log out
| 1
|
11,016
| 13,803,061,544
|
IssuesEvent
|
2020-10-11 00:27:07
|
SpencerTSterling/RatHutWebsite
|
https://api.github.com/repos/SpencerTSterling/RatHutWebsite
|
closed
|
Set up Continuous Integration with GitHub Actions
|
development process
|
GitHub actions should be used to build the project on each commit
|
1.0
|
Set up Continuous Integration with GitHub Actions - GitHub actions should be used to build the project on each commit
|
process
|
set up continuous integration with github actions github actions should be used to build the project on each commit
| 1
|
1,276
| 3,811,709,600
|
IssuesEvent
|
2016-03-27 01:23:08
|
dotnet/corefx
|
https://api.github.com/repos/dotnet/corefx
|
opened
|
ServiceControllerTests.ControlCapabilities failed in CI on Win7
|
System.ServiceProcess
|
http://dotnet-ci.cloudapp.net/job/dotnet_corefx/job/outerloop_win7_debug/136/consoleFull
```
09:29:26 System.ServiceProcess.Tests.ServiceControllerTests.ControlCapabilities [FAIL]
09:29:26 Assert.True() Failure
09:29:26 Expected: True
09:29:26 Actual: False
09:29:26 Stack Trace:
09:29:26 at System.ServiceProcess.Tests.ServiceControllerTests.ControlCapabilities()
09:29:27 System.Threading.Overlapped -> d:\j\workspace\outerloop_win---9c9e7d59\bin\ref\System.Th
```
|
1.0
|
ServiceControllerTests.ControlCapabilities failed in CI on Win7 - http://dotnet-ci.cloudapp.net/job/dotnet_corefx/job/outerloop_win7_debug/136/consoleFull
```
09:29:26 System.ServiceProcess.Tests.ServiceControllerTests.ControlCapabilities [FAIL]
09:29:26 Assert.True() Failure
09:29:26 Expected: True
09:29:26 Actual: False
09:29:26 Stack Trace:
09:29:26 at System.ServiceProcess.Tests.ServiceControllerTests.ControlCapabilities()
09:29:27 System.Threading.Overlapped -> d:\j\workspace\outerloop_win---9c9e7d59\bin\ref\System.Th
```
|
process
|
servicecontrollertests controlcapabilities failed in ci on system serviceprocess tests servicecontrollertests controlcapabilities assert true failure expected true actual false stack trace at system serviceprocess tests servicecontrollertests controlcapabilities system threading overlapped d j workspace outerloop win bin ref system th
| 1
|
5,574
| 8,409,145,075
|
IssuesEvent
|
2018-10-12 06:00:12
|
rancher/rancher
|
https://api.github.com/repos/rancher/rancher
|
closed
|
Add support for security_opt: - "credentialspec=file://gmsaDocker.json"
|
area/agent area/windows kind/bug process/cherry-pick version/1.6
|
**Rancher Versions:**
Server: 1.5.1
**Docker Version:**
17.03.0-ce
**OS and where are the hosts located? (cloud, bare metal, etc):**
windows server 2016 datacenter on hyper-v vm with nested virtualization enabled
**Setup Details: (single node rancher vs. HA rancher, internal DB vs. external DB)**
2 rancher manager nodes, currently only 1 is in use, haproxy, 3 mariadb-s with galera
**Environment Type: (Cattle/Kubernetes/Swarm/Mesos)**
windows
**Steps to Reproduce:**
1) create gmsa Active Directory account for the container
https://docs.microsoft.com/en-us/virtualization/windowscontainers/manage-containers/manage-serviceaccounts
2) export credentials to json file
3) run windows container from rancher UI(or from rancher-cli) with docker-compose.yml
**Results:**
No success. If it will be done from the windows node(where rancher agent is installed) using docker CLI it'll be successful. Reason - that yml file with configuration of the rancher service has a string **credentialspec=file://gmsaDocker.json**, but this file will not be transmitted to server, even if is located in the same folder as the rancher-cli executable
version: '2'
services:
test-express-23-dc:
image: "microsoft/windowsservercore"
network_mode: "transparent"
security_opt:
**- "credentialspec=file://gmsaDocker.json"**
does not work because file is not transmitted to docker service
That parameter works with docker cli only if there is a file gmsaDocker.json that is located in c:\ProgramData\Docker\Credentialspec
**Suggested Improvements:**
some way to provide this file from rancher-cli to rancher-server and to rancher-agent
may be manually put that file to some location on the target machine, for example c:\programmdata\docker
|
1.0
|
Add support for security_opt: - "credentialspec=file://gmsaDocker.json" - **Rancher Versions:**
Server: 1.5.1
**Docker Version:**
17.03.0-ce
**OS and where are the hosts located? (cloud, bare metal, etc):**
windows server 2016 datacenter on hyper-v vm with nested virtualization enabled
**Setup Details: (single node rancher vs. HA rancher, internal DB vs. external DB)**
2 rancher manager nodes, currently only 1 is in use, haproxy, 3 mariadb-s with galera
**Environment Type: (Cattle/Kubernetes/Swarm/Mesos)**
windows
**Steps to Reproduce:**
1) create gmsa Active Directory account for the container
https://docs.microsoft.com/en-us/virtualization/windowscontainers/manage-containers/manage-serviceaccounts
2) export credentials to json file
3) run windows container from rancher UI(or from rancher-cli) with docker-compose.yml
**Results:**
No success. If it will be done from the windows node(where rancher agent is installed) using docker CLI it'll be successful. Reason - that yml file with configuration of the rancher service has a string **credentialspec=file://gmsaDocker.json**, but this file will not be transmitted to server, even if is located in the same folder as the rancher-cli executable
version: '2'
services:
test-express-23-dc:
image: "microsoft/windowsservercore"
network_mode: "transparent"
security_opt:
**- "credentialspec=file://gmsaDocker.json"**
does not work because file is not transmitted to docker service
That parameter works with docker cli only if there is a file gmsaDocker.json that is located in c:\ProgramData\Docker\Credentialspec
**Suggested Improvements:**
some way to provide this file from rancher-cli to rancher-server and to rancher-agent
may be manually put that file to some location on the target machine, for example c:\programmdata\docker
|
process
|
add support for security opt credentialspec file gmsadocker json rancher versions server docker version ce os and where are the hosts located cloud bare metal etc windows server datacenter on hyper v vm with nested virtualization enabled setup details single node rancher vs ha rancher internal db vs external db rancher manager nodes currently only is in use haproxy mariadb s with galera environment type cattle kubernetes swarm mesos windows steps to reproduce create gmsa active directory account for the container export credentials to json file run windows container from rancher ui or from rancher cli with docker compose yml results no success if it will be done from the windows node where rancher agent is installed using docker cli it ll be successful reason that yml file with configuration of the rancher service has a string credentialspec file gmsadocker json but this file will not be transmitted to server even if is located in the same folder as the rancher cli executable version services test express dc image microsoft windowsservercore network mode transparent security opt credentialspec file gmsadocker json does not work because file is not transmitted to docker service that parameter works with docker cli only if there is a file gmsadocker json that is located in c programdata docker credentialspec suggested improvements some way to provide this file from rancher cli to rancher server and to rancher agent may be manually put that file to some location on the target machine for example c programmdata docker
| 1
|
822,275
| 30,862,425,166
|
IssuesEvent
|
2023-08-03 05:01:38
|
WordPress/openverse
|
https://api.github.com/repos/WordPress/openverse
|
opened
|
Component: `VTag`
|
good first issue help wanted 🟨 priority: medium 🌟 goal: addition 🕹 aspect: interface 🧱 stack: frontend
|
## Description
<!-- Describe the component, including different states. Do not include screenshots. -->
The new `VTag` component will be used for clickable tags. The current, non-clickable, `VMediaTag` component will be replaced with it.
To create a new component, run `just frontend/run create:component VTag`. This will run the component generator and will create the new component file, a Storybook story, unit and Playwright tests. The unit and playwright tests can be deleted.
## API
<!-- Tentatively specify the props, state and emitted events of the component. -->
### Props
<!-- Describe possible props and their types; or delete the section if the component takes no props. -->
| Name | Type | Description |
|-|-|-|
| title | string | The tag text |
| href | string | The url the tag should link to |
## Code samples
<!-- Share pseudocode templates or high-level implementation code; or delete the section entirely. -->
```
<VButton
as="VLink"
size="small"
variant="filled-gray"
class="label-bold"
:href="href"
>{{ title }}</VButton>
```
<!-- Place a link to the Figma node of the component from the Design Library: https://www.figma.com/file/GIIQ4sDbaToCfFQyKMvzr8/Openverse-Design-Library -->
- **Figma:** https://www.figma.com/file/niWnCgB7K0Y4e4mgxMrnRC/Additional-search-views?type=design&node-id=1263-62874&mode=design&t=WRJSFZNUbmpYERx8-4
|
1.0
|
Component: `VTag` - ## Description
<!-- Describe the component, including different states. Do not include screenshots. -->
The new `VTag` component will be used for clickable tags. The current, non-clickable, `VMediaTag` component will be replaced with it.
To create a new component, run `just frontend/run create:component VTag`. This will run the component generator and will create the new component file, a Storybook story, unit and Playwright tests. The unit and playwright tests can be deleted.
## API
<!-- Tentatively specify the props, state and emitted events of the component. -->
### Props
<!-- Describe possible props and their types; or delete the section if the component takes no props. -->
| Name | Type | Description |
|-|-|-|
| title | string | The tag text |
| href | string | The url the tag should link to |
## Code samples
<!-- Share pseudocode templates or high-level implementation code; or delete the section entirely. -->
```
<VButton
as="VLink"
size="small"
variant="filled-gray"
class="label-bold"
:href="href"
>{{ title }}</VButton>
```
<!-- Place a link to the Figma node of the component from the Design Library: https://www.figma.com/file/GIIQ4sDbaToCfFQyKMvzr8/Openverse-Design-Library -->
- **Figma:** https://www.figma.com/file/niWnCgB7K0Y4e4mgxMrnRC/Additional-search-views?type=design&node-id=1263-62874&mode=design&t=WRJSFZNUbmpYERx8-4
|
non_process
|
component vtag description the new vtag component will be used for clickable tags the current non clickable vmediatag component will be replaced with it to create a new component run just frontend run create component vtag this will run the component generator and will create the new component file a storybook story unit and playwright tests the unit and playwright tests can be deleted api props name type description title string the tag text href string the url the tag should link to code samples vbutton as vlink size small variant filled gray class label bold href href title figma
| 0
|
576,457
| 17,087,317,763
|
IssuesEvent
|
2021-07-08 13:26:10
|
ParamagicDev/mrujs
|
https://api.github.com/repos/ParamagicDev/mrujs
|
opened
|
feat: Allow for end users to perform manual page caching
|
Priority: Normal Project Size: Small Type: Feature
|
#### Describe Problem
Users cannot easily prefetch links
#### Suggest Solution
Expose the navigationAdapter's functions to fetch the cache.
|
1.0
|
feat: Allow for end users to perform manual page caching - #### Describe Problem
Users cannot easily prefetch links
#### Suggest Solution
Expose the navigationAdapter's functions to fetch the cache.
|
non_process
|
feat allow for end users to perform manual page caching describe problem users cannot easily prefetch links suggest solution expose the navigationadapter s functions to fetch the cache
| 0
|
746,973
| 26,052,220,803
|
IssuesEvent
|
2022-12-22 20:01:11
|
evanshunt/derekstrap
|
https://api.github.com/repos/evanshunt/derekstrap
|
closed
|
1.0 ideas - Breakpoint method onBreakpoint();
|
enhancement Medium priority in progress
|
The `window.breakpointChanged` method can already be used to detect crossing a specific breakpoint threshold, but a simpler helper method that allowed you to pass in a breakpoint name and callbacks for what happens when the threshold is crossed would be handy.
|
1.0
|
1.0 ideas - Breakpoint method onBreakpoint(); - The `window.breakpointChanged` method can already be used to detect crossing a specific breakpoint threshold, but a simpler helper method that allowed you to pass in a breakpoint name and callbacks for what happens when the threshold is crossed would be handy.
|
non_process
|
ideas breakpoint method onbreakpoint the window breakpointchanged method can already be used to detect crossing a specific breakpoint threshold but a simpler helper method that allowed you to pass in a breakpoint name and callbacks for what happens when the threshold is crossed would be handy
| 0
|
41,446
| 8,973,322,044
|
IssuesEvent
|
2019-01-29 20:42:16
|
Foggalong/hardcode-fixer
|
https://api.github.com/repos/Foggalong/hardcode-fixer
|
closed
|
Return to castle wolfenstein SP
|
hardcoded app
|
| Application name | Icon name | Desktop file name |
| --- | :-- | :-- |
| Return to castle wolfenstein SP | `Icon=/usr/share/pixmaps/WolfSP.xpm` | `wolfsp.desktop` |
|
1.0
|
Return to castle wolfenstein SP - | Application name | Icon name | Desktop file name |
| --- | :-- | :-- |
| Return to castle wolfenstein SP | `Icon=/usr/share/pixmaps/WolfSP.xpm` | `wolfsp.desktop` |
|
non_process
|
return to castle wolfenstein sp application name icon name desktop file name return to castle wolfenstein sp icon usr share pixmaps wolfsp xpm wolfsp desktop
| 0
|
4,894
| 2,760,951,954
|
IssuesEvent
|
2015-04-28 14:56:16
|
mozilla/webmaker-app
|
https://api.github.com/repos/mozilla/webmaker-app
|
closed
|
UI - Bottom navigation components
|
design in progress
|
Finalize UI design for the bottom navigation components.
Go to InVision to leave feedback: http://invis.io/BK2M18Q28
|
1.0
|
UI - Bottom navigation components - Finalize UI design for the bottom navigation components.
Go to InVision to leave feedback: http://invis.io/BK2M18Q28
|
non_process
|
ui bottom navigation components finalize ui design for the bottom navigation components go to invision to leave feedback
| 0
|
173,981
| 27,554,871,758
|
IssuesEvent
|
2023-03-07 17:11:21
|
blockprotocol/blockprotocol
|
https://api.github.com/repos/blockprotocol/blockprotocol
|
opened
|
Set up visual regression testing with Playwright
|
area: design system
|
Look to leverage our existing BrowserStack subscription with something like BrowserStack Percy's integration w/ Playwright: https://docs.percy.io/docs/playwright
|
1.0
|
Set up visual regression testing with Playwright - Look to leverage our existing BrowserStack subscription with something like BrowserStack Percy's integration w/ Playwright: https://docs.percy.io/docs/playwright
|
non_process
|
set up visual regression testing with playwright look to leverage our existing browserstack subscription with something like browserstack percy s integration w playwright
| 0
|
10,931
| 13,745,297,207
|
IssuesEvent
|
2020-10-06 02:26:59
|
Alphadrake86/CraftsMadeByHand
|
https://api.github.com/repos/Alphadrake86/CraftsMadeByHand
|
closed
|
Set up CI pipeline
|
Dev Processes
|
Set up Continuous Integration pipeline to ensure all incoming commits compile
|
1.0
|
Set up CI pipeline - Set up Continuous Integration pipeline to ensure all incoming commits compile
|
process
|
set up ci pipeline set up continuous integration pipeline to ensure all incoming commits compile
| 1
|
49,263
| 13,186,563,223
|
IssuesEvent
|
2020-08-13 00:34:17
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
opened
|
clsim - hobo clsim build chokes doc build (Trac #1077)
|
Incomplete Migration Migrated from Trac combo simulation defect
|
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1077">https://code.icecube.wisc.edu/ticket/1077</a>, reported by nega and owned by claudio.kopper</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:15:08",
"description": "clsim's dependency list should probably AND'd, not OR'd. \n\n{{{\n-- + clsim\n-- +-- python [symlinks] \n-- +-- Geant4 or OpenCL is not installed on your system. clsim will fail if it is not used with parameterizations. \n-- +-- numpy support (for tabulator) \n-- +-- safeprimes_base32.gz already downloaded \n-- +-- gmp support (make_safeprimes utility) \n-- +-- clsim-pybindings \n}}}\n\ngives:\n{{{\n[ 62%] Generating html from icetray-inspect of clsim\nIgnoring 'clsim': dlopen() dynamic loading error: /home/nega/i3/combo/build/lib/libclsim.so: undefined symbol: clCreateSubDevices/home/nega/i3/combo/build/CMakeFiles/clsim-inspection.xml:4: parser error : Opening and ending tag mismatch: project line 3 and icetray-inspect\n</icetray-inspect>\n ^\n/home/nega/i3/combo/build/CMakeFiles/clsim-inspection.xml:5: parser error : Premature end of data in tag icetray-inspect line 2\n\n^\nunable to parse /home/nega/i3/combo/build/CMakeFiles/clsim-inspection.xml\nmake[3]: *** [clsim/CMakeFiles/clsim-clsim-inspect] Error 6\nmake[2]: *** [clsim/CMakeFiles/clsim-clsim-inspect.dir/all] Error 2\nmake[1]: *** [CMakeFiles/docs.dir/rule] Error 2\nmake: *** [docs] Error 2\n \n~/i3/combo/build 1m 53s\n\u276f \n}}}",
"reporter": "nega",
"cc": "",
"resolution": "worksforme",
"_ts": "1550067308113782",
"component": "combo simulation",
"summary": "clsim - hobo clsim build chokes doc build",
"priority": "normal",
"keywords": "clsim docs",
"time": "2015-07-30T04:18:17",
"milestone": "",
"owner": "claudio.kopper",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
clsim - hobo clsim build chokes doc build (Trac #1077) - <details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/ticket/1077">https://code.icecube.wisc.edu/ticket/1077</a>, reported by nega and owned by claudio.kopper</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:15:08",
"description": "clsim's dependency list should probably AND'd, not OR'd. \n\n{{{\n-- + clsim\n-- +-- python [symlinks] \n-- +-- Geant4 or OpenCL is not installed on your system. clsim will fail if it is not used with parameterizations. \n-- +-- numpy support (for tabulator) \n-- +-- safeprimes_base32.gz already downloaded \n-- +-- gmp support (make_safeprimes utility) \n-- +-- clsim-pybindings \n}}}\n\ngives:\n{{{\n[ 62%] Generating html from icetray-inspect of clsim\nIgnoring 'clsim': dlopen() dynamic loading error: /home/nega/i3/combo/build/lib/libclsim.so: undefined symbol: clCreateSubDevices/home/nega/i3/combo/build/CMakeFiles/clsim-inspection.xml:4: parser error : Opening and ending tag mismatch: project line 3 and icetray-inspect\n</icetray-inspect>\n ^\n/home/nega/i3/combo/build/CMakeFiles/clsim-inspection.xml:5: parser error : Premature end of data in tag icetray-inspect line 2\n\n^\nunable to parse /home/nega/i3/combo/build/CMakeFiles/clsim-inspection.xml\nmake[3]: *** [clsim/CMakeFiles/clsim-clsim-inspect] Error 6\nmake[2]: *** [clsim/CMakeFiles/clsim-clsim-inspect.dir/all] Error 2\nmake[1]: *** [CMakeFiles/docs.dir/rule] Error 2\nmake: *** [docs] Error 2\n \n~/i3/combo/build 1m 53s\n\u276f \n}}}",
"reporter": "nega",
"cc": "",
"resolution": "worksforme",
"_ts": "1550067308113782",
"component": "combo simulation",
"summary": "clsim - hobo clsim build chokes doc build",
"priority": "normal",
"keywords": "clsim docs",
"time": "2015-07-30T04:18:17",
"milestone": "",
"owner": "claudio.kopper",
"type": "defect"
}
```
</p>
</details>
|
non_process
|
clsim hobo clsim build chokes doc build trac migrated from json status closed changetime description clsim s dependency list should probably and d not or d n n n clsim n python n or opencl is not installed on your system clsim will fail if it is not used with parameterizations n numpy support for tabulator n safeprimes gz already downloaded n gmp support make safeprimes utility n clsim pybindings n n ngives n n generating html from icetray inspect of clsim nignoring clsim dlopen dynamic loading error home nega combo build lib libclsim so undefined symbol clcreatesubdevices home nega combo build cmakefiles clsim inspection xml parser error opening and ending tag mismatch project line and icetray inspect n n n home nega combo build cmakefiles clsim inspection xml parser error premature end of data in tag icetray inspect line n n nunable to parse home nega combo build cmakefiles clsim inspection xml nmake error nmake error nmake error nmake error n n combo build n n reporter nega cc resolution worksforme ts component combo simulation summary clsim hobo clsim build chokes doc build priority normal keywords clsim docs time milestone owner claudio kopper type defect
| 0
|
18,978
| 24,966,042,743
|
IssuesEvent
|
2022-11-01 19:28:28
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
qgis_process temporary files are not cleaned up
|
Processing Bug
|
### What is the bug or the crash?
On each call to `qgis_process`, two new files are created in the user's temporary directory, but they are not removed afterwards:
- an empty file `QGIS3.xxxxxx`
- a file in a subdirectory: `QGIS3-yyyyyy/zzzzzz_styles.db`. This is a SQLite database with 14 tables having 0 rows.
The xxx/yyy/zzz strings seem random.
With several tested algorithms, these files all appear to remain empty / with 0 rows. If this is always the case, then perhaps the files shouldn't be there?
When checking the R package qgisprocess, this behaviour yields a note about remaining detritus (https://github.com/paleolimbot/qgisprocess/issues/62).
### Steps to reproduce the issue
In a Linux shell:
```bash
$ find /tmp -type f \( -name "QGIS*" -o -name "*_styles.db" \)
$
$ qgis_process &> /dev/null
$
$ find /tmp -type f \( -name "QGIS*" -o -name "*_styles.db" \)
/tmp/QGIS3-UJyKwM/KHMJUP_styles.db
/tmp/QGIS3.BHULds
$
$ qgis_process &> /dev/null
$
$ find /tmp -type f \( -name "QGIS*" -o -name "*_styles.db" \)
/tmp/QGIS3-UJyKwM/KHMJUP_styles.db
/tmp/QGIS3.BHULds
/tmp/QGIS3.YBTuhF
/tmp/QGIS3-Ovuiht/hHUlfq_styles.db
```
### Versions
<!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.0//EN" "http://www.w3.org/TR/REC-html40/strict.dtd">
<html><head><meta http-equiv="Content-Type" content="text/html; charset=utf-8" /></head><body>
QGIS version | 3.26.3-Buenos Aires | QGIS code revision | 65e4edfdad
-- | -- | -- | --
Qt version | 5.12.8
Python version | 3.8.10
GDAL/OGR version | 3.4.3
PROJ version | 8.2.0
EPSG Registry database version | v10.038 (2021-10-21)
GEOS version | 3.10.2-CAPI-1.16.0
SQLite version | 3.31.1
PDAL version | 2.2.0
PostgreSQL client version | 12.12 (Ubuntu 12.12-0ubuntu0.20.04.1)
SpatiaLite version | 5.0.1
QWT version | 6.1.4
QScintilla2 version | 2.11.2
OS version | Linux Mint 20
| | |
Active Python plugins
geopunt4Qgis | 2.2.4
ViewshedAnalysis | 1.7
cartography_tools | 1.2.1
quick_map_services | 0.19.27
grassprovider | 2.12.99
processing | 2.12.99
sagaprovider | 2.12.99
db_manager | 0.1.20
MetaSearch | 0.3.6
otbprovider | 2.12.99
</body></html>
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [ ] I tried with a new QGIS profile
### Additional context
_No response_
|
1.0
|
qgis_process temporary files are not cleaned up - ### What is the bug or the crash?
On each call to `qgis_process`, two new files are created in the user's temporary directory, but they are not removed afterwards:
- an empty file `QGIS3.xxxxxx`
- a file in a subdirectory: `QGIS3-yyyyyy/zzzzzz_styles.db`. This is a SQLite database with 14 tables having 0 rows.
The xxx/yyy/zzz strings seem random.
With several tested algorithms, these files all appear to remain empty / with 0 rows. If this is always the case, then perhaps the files shouldn't be there?
When checking the R package qgisprocess, this behaviour yields a note about remaining detritus (https://github.com/paleolimbot/qgisprocess/issues/62).
### Steps to reproduce the issue
In a Linux shell:
```bash
$ find /tmp -type f \( -name "QGIS*" -o -name "*_styles.db" \)
$
$ qgis_process &> /dev/null
$
$ find /tmp -type f \( -name "QGIS*" -o -name "*_styles.db" \)
/tmp/QGIS3-UJyKwM/KHMJUP_styles.db
/tmp/QGIS3.BHULds
$
$ qgis_process &> /dev/null
$
$ find /tmp -type f \( -name "QGIS*" -o -name "*_styles.db" \)
/tmp/QGIS3-UJyKwM/KHMJUP_styles.db
/tmp/QGIS3.BHULds
/tmp/QGIS3.YBTuhF
/tmp/QGIS3-Ovuiht/hHUlfq_styles.db
```
### Versions
<!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.0//EN" "http://www.w3.org/TR/REC-html40/strict.dtd">
<html><head><meta http-equiv="Content-Type" content="text/html; charset=utf-8" /></head><body>
QGIS version | 3.26.3-Buenos Aires | QGIS code revision | 65e4edfdad
-- | -- | -- | --
Qt version | 5.12.8
Python version | 3.8.10
GDAL/OGR version | 3.4.3
PROJ version | 8.2.0
EPSG Registry database version | v10.038 (2021-10-21)
GEOS version | 3.10.2-CAPI-1.16.0
SQLite version | 3.31.1
PDAL version | 2.2.0
PostgreSQL client version | 12.12 (Ubuntu 12.12-0ubuntu0.20.04.1)
SpatiaLite version | 5.0.1
QWT version | 6.1.4
QScintilla2 version | 2.11.2
OS version | Linux Mint 20
| | |
Active Python plugins
geopunt4Qgis | 2.2.4
ViewshedAnalysis | 1.7
cartography_tools | 1.2.1
quick_map_services | 0.19.27
grassprovider | 2.12.99
processing | 2.12.99
sagaprovider | 2.12.99
db_manager | 0.1.20
MetaSearch | 0.3.6
otbprovider | 2.12.99
</body></html>
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [ ] I tried with a new QGIS profile
### Additional context
_No response_
|
process
|
qgis process temporary files are not cleaned up what is the bug or the crash on each call to qgis process two new files are created in the user s temporary directory but they are not removed afterwards an empty file xxxxxx a file in a subdirectory yyyyyy zzzzzz styles db this is a sqlite database with tables having rows the xxx yyy zzz strings seem random with several tested algorithms these files all appear to remain empty with rows if this is always the case then perhaps the files shouldn t be there when checking the r package qgisprocess this behaviour yields a note about remaining detritus steps to reproduce the issue in a linux shell bash find tmp type f name qgis o name styles db qgis process dev null find tmp type f name qgis o name styles db tmp ujykwm khmjup styles db tmp bhulds qgis process dev null find tmp type f name qgis o name styles db tmp ujykwm khmjup styles db tmp bhulds tmp ybtuhf tmp ovuiht hhulfq styles db versions doctype html public dtd html en qgis version buenos aires qgis code revision qt version python version gdal ogr version proj version epsg registry database version geos version capi sqlite version pdal version postgresql client version ubuntu spatialite version qwt version version os version linux mint active python plugins viewshedanalysis cartography tools quick map services grassprovider processing sagaprovider db manager metasearch otbprovider supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context no response
| 1
|
699,358
| 24,013,874,252
|
IssuesEvent
|
2022-09-14 21:37:41
|
tl-its-umich-edu/my-learning-analytics
|
https://api.github.com/repos/tl-its-umich-edu/my-learning-analytics
|
closed
|
Vulnerability: CVE-2015-20107, Python
|
⚠️ priority ☠ vulnerability
|
From [vulnerabilities](https://docs.google.com/spreadsheets/d/12cyQTkIxCxwP0lAi58eFvznBcKDO3yB3mKdSenGkBOM) spreadsheet, based on Unizin analysis:
CVE ID | Effective Severity | Severity | Impacted Image | Vulnerable Package | Remediated Package | URL
-- | -- | -- | -- | -- | -- | --
CVE-2015-20107 | CRITICAL | CRITICAL | gcr.io/unizin-core/myla:2022.01.01 | python3.9/3.9.2-1 | python3.9/MAXIMUM | https://security-tracker.debian.org/tracker/CVE-2015-20107
|
1.0
|
Vulnerability: CVE-2015-20107, Python - From [vulnerabilities](https://docs.google.com/spreadsheets/d/12cyQTkIxCxwP0lAi58eFvznBcKDO3yB3mKdSenGkBOM) spreadsheet, based on Unizin analysis:
CVE ID | Effective Severity | Severity | Impacted Image | Vulnerable Package | Remediated Package | URL
-- | -- | -- | -- | -- | -- | --
CVE-2015-20107 | CRITICAL | CRITICAL | gcr.io/unizin-core/myla:2022.01.01 | python3.9/3.9.2-1 | python3.9/MAXIMUM | https://security-tracker.debian.org/tracker/CVE-2015-20107
|
non_process
|
vulnerability cve python from spreadsheet based on unizin analysis cve id effective severity severity impacted image vulnerable package remediated package url cve critical critical gcr io unizin core myla maximum
| 0
|
198,566
| 22,659,667,161
|
IssuesEvent
|
2022-07-02 01:15:21
|
snykiotcubedev/arangodb-3.7.6
|
https://api.github.com/repos/snykiotcubedev/arangodb-3.7.6
|
reopened
|
CVE-2021-23382 (Medium) detected in postcss-7.0.32.tgz
|
security vulnerability
|
## CVE-2021-23382 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>postcss-7.0.32.tgz</b></p></summary>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-7.0.32.tgz">https://registry.npmjs.org/postcss/-/postcss-7.0.32.tgz</a></p>
<p>
Dependency Hierarchy:
- css-loader-2.1.1.tgz (Root Library)
- :x: **postcss-7.0.32.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/snykiotcubedev/arangodb-3.7.6/commit/fce8f85f1c2f070c8e6a8e76d17210a2117d3833">fce8f85f1c2f070c8e6a8e76d17210a2117d3833</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package postcss before 8.2.13 are vulnerable to Regular Expression Denial of Service (ReDoS) via getAnnotationURL() and loadAnnotation() in lib/previous-map.js. The vulnerable regexes are caused mainly by the sub-pattern \/\*\s* sourceMappingURL=(.*).
<p>Publish Date: 2021-04-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23382>CVE-2021-23382</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23382">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23382</a></p>
<p>Release Date: 2021-04-26</p>
<p>Fix Resolution (postcss): 7.0.36</p>
<p>Direct dependency fix Resolution (css-loader): 3.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-23382 (Medium) detected in postcss-7.0.32.tgz - ## CVE-2021-23382 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>postcss-7.0.32.tgz</b></p></summary>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-7.0.32.tgz">https://registry.npmjs.org/postcss/-/postcss-7.0.32.tgz</a></p>
<p>
Dependency Hierarchy:
- css-loader-2.1.1.tgz (Root Library)
- :x: **postcss-7.0.32.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/snykiotcubedev/arangodb-3.7.6/commit/fce8f85f1c2f070c8e6a8e76d17210a2117d3833">fce8f85f1c2f070c8e6a8e76d17210a2117d3833</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package postcss before 8.2.13 are vulnerable to Regular Expression Denial of Service (ReDoS) via getAnnotationURL() and loadAnnotation() in lib/previous-map.js. The vulnerable regexes are caused mainly by the sub-pattern \/\*\s* sourceMappingURL=(.*).
<p>Publish Date: 2021-04-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23382>CVE-2021-23382</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23382">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23382</a></p>
<p>Release Date: 2021-04-26</p>
<p>Fix Resolution (postcss): 7.0.36</p>
<p>Direct dependency fix Resolution (css-loader): 3.0.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in postcss tgz cve medium severity vulnerability vulnerable library postcss tgz tool for transforming styles with js plugins library home page a href dependency hierarchy css loader tgz root library x postcss tgz vulnerable library found in head commit a href found in base branch main vulnerability details the package postcss before are vulnerable to regular expression denial of service redos via getannotationurl and loadannotation in lib previous map js the vulnerable regexes are caused mainly by the sub pattern s sourcemappingurl publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution postcss direct dependency fix resolution css loader step up your open source security game with mend
| 0
|
17,492
| 23,305,507,934
|
IssuesEvent
|
2022-08-07 23:50:04
|
lynnandtonic/nestflix.fun
|
https://api.github.com/repos/lynnandtonic/nestflix.fun
|
closed
|
Add Lil Ark Angel “Cats Don’t Dance” (Screenshots and Title Screen Added)
|
suggested title in process rental
|
Please add as much of the following info as you can:
Title: Lil Ark Angel
Type (film/tv show): Film - Golden Age musical
Film or show in which it appears: Cats Don't Dance
Is the parent film/show streaming anywhere? Yes - Amazon Prime
About when in the parent film/show does it appear? probably about the first third of the film
Actual footage of the film/show can be seen (yes/no)? Yes. Here is the musical number "Little Boat on the Sea": https://www.youtube.com/watch?v=bShn4m6I49k
Cast: Darla Dimple
Director: Flanagan
Production Company: Mammoth Pictures
Quote: I built a little boat as cute as it can be, and I put all the animals two-by-two in my little boat on the sea.
|
1.0
|
Add Lil Ark Angel “Cats Don’t Dance” (Screenshots and Title Screen Added) - Please add as much of the following info as you can:
Title: Lil Ark Angel
Type (film/tv show): Film - Golden Age musical
Film or show in which it appears: Cats Don't Dance
Is the parent film/show streaming anywhere? Yes - Amazon Prime
About when in the parent film/show does it appear? probably about the first third of the film
Actual footage of the film/show can be seen (yes/no)? Yes. Here is the musical number "Little Boat on the Sea": https://www.youtube.com/watch?v=bShn4m6I49k
Cast: Darla Dimple
Director: Flanagan
Production Company: Mammoth Pictures
Quote: I built a little boat as cute as it can be, and I put all the animals two-by-two in my little boat on the sea.
|
process
|
add lil ark angel “cats don’t dance” screenshots and title screen added please add as much of the following info as you can title lil ark angel type film tv show film golden age musical film or show in which it appears cats don t dance is the parent film show streaming anywhere yes amazon prime about when in the parent film show does it appear probably about the first third of the film actual footage of the film show can be seen yes no yes here is the musical number little boat on the sea cast darla dimple director flanagan production company mammoth pictures quote i built a little boat as cute as it can be and i put all the animals two by two in my little boat on the sea
| 1
|
11,413
| 14,242,136,981
|
IssuesEvent
|
2020-11-19 01:02:44
|
googleapis/java-speech
|
https://api.github.com/repos/googleapis/java-speech
|
closed
|
longRunningRecognizeAsync in asyncRecognizeGcs doesn't work with long audio files.
|
api: speech type: process
|
While troubleshooting this [issue](https://github.com/GoogleCloudPlatform/php-docs-samples/issues/1094) I discovered that the ```longRunningRecognizeAsync ```in this [sample](https://github.com/googleapis/java-speech/blob/master/samples/snippets/src/main/java/com/example/speech/Recognize.java) presents a similar behavior as the original issue I was looking into.
To repro:
Use the ```asyncRecognizeGcs ``` from the [samples](https://github.com/googleapis/java-speech/blob/master/samples/snippets/src/main/java/com/example/speech/Recognize.java) and feed it a long gcs audio file (at least 30 minutes long) and run it.
The process will run for a while to render no results.
Will see a "Task was cancelled." message.
|
1.0
|
longRunningRecognizeAsync in asyncRecognizeGcs doesn't work with long audio files. - While troubleshooting this [issue](https://github.com/GoogleCloudPlatform/php-docs-samples/issues/1094) I discovered that the ```longRunningRecognizeAsync ```in this [sample](https://github.com/googleapis/java-speech/blob/master/samples/snippets/src/main/java/com/example/speech/Recognize.java) presents a similar behavior as the original issue I was looking into.
To repro:
Use the ```asyncRecognizeGcs ``` from the [samples](https://github.com/googleapis/java-speech/blob/master/samples/snippets/src/main/java/com/example/speech/Recognize.java) and feed it a long gcs audio file (at least 30 minutes long) and run it.
The process will run for a while to render no results.
Will see a "Task was cancelled." message.
|
process
|
longrunningrecognizeasync in asyncrecognizegcs doesn t work with long audio files while troubleshooting this i discovered that the longrunningrecognizeasync in this presents a similar behavior as the original issue i was looking into to repro use the asyncrecognizegcs from the and feed it a long gcs audio file at least minutes long and run it the process will run for a while to render no results will see a task was cancelled message
| 1
|
9,858
| 12,857,494,627
|
IssuesEvent
|
2020-07-09 09:23:51
|
zammad/zammad
|
https://api.github.com/repos/zammad/zammad
|
opened
|
ServiceNow mails from other service providers are not detected
|
bug mail processing
|
### Infos:
* Used Zammad version: 3.4
* Installation method (source, package, ..): all
* Operating system: all
* Database + version: all
* Elasticsearch version: all
* Browser + version: all
### Expected behavior:
* ServiceNow mails should be detected based on the mail headers, instead of the sender address service-now.com
### Actual behavior:
* ServiceNow mails from different sender addresses than service-now.com are not detected.
### Steps to reproduce the behavior:
-
|
1.0
|
ServiceNow mails from other service providers are not detected -
### Infos:
* Used Zammad version: 3.4
* Installation method (source, package, ..): all
* Operating system: all
* Database + version: all
* Elasticsearch version: all
* Browser + version: all
### Expected behavior:
* ServiceNow mails should be detected based on the mail headers, instead of the sender address service-now.com
### Actual behavior:
* ServiceNow mails from different sender addresses than service-now.com are not detected.
### Steps to reproduce the behavior:
-
|
process
|
servicenow mails from other service providers are not detected infos used zammad version installation method source package all operating system all database version all elasticsearch version all browser version all expected behavior servicenow mails should be detected based on the mail headers instead of the sender address service now com actual behavior servicenow mails from different sender addresses than service now com are not detected steps to reproduce the behavior
| 1
|
11,972
| 14,737,010,373
|
IssuesEvent
|
2021-01-07 00:36:50
|
kdjstudios/SABillingGitlab
|
https://api.github.com/repos/kdjstudios/SABillingGitlab
|
closed
|
not able to make cc charges from billing site
|
anc-external anc-process anp-important anp-urgent ant-bug ant-support
|
In GitLab by @kdjstudios on Apr 9, 2018, 15:48
**Submitted by:** Gaylan Garrett <gaylan@keenercom.net>
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-04-09-52796/conversation
**Server:** Hosted
**Client/Site:** Keener
**Account:** NA
**Issue:**
I just wanted to make you aware that I am not able to do a credit card charge from the site. I have to go to authorize.net and make the charge and then return to SA billing to post the charge.
|
1.0
|
not able to make cc charges from billing site - In GitLab by @kdjstudios on Apr 9, 2018, 15:48
**Submitted by:** Gaylan Garrett <gaylan@keenercom.net>
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-04-09-52796/conversation
**Server:** Hosted
**Client/Site:** Keener
**Account:** NA
**Issue:**
I just wanted to make you aware that I am not able to do a credit card charge from the site. I have to go to authorize.net and make the charge and then return to SA billing to post the charge.
|
process
|
not able to make cc charges from billing site in gitlab by kdjstudios on apr submitted by gaylan garrett helpdesk server hosted client site keener account na issue i just wanted to make you aware that i am not able to do a credit card charge from the site i have to go to authorize net and make the charge and then return to sa billing to post the charge
| 1
|
10,318
| 13,160,837,615
|
IssuesEvent
|
2020-08-10 18:21:52
|
GoogleCloudPlatform/stackdriver-sandbox
|
https://api.github.com/repos/GoogleCloudPlatform/stackdriver-sandbox
|
closed
|
Move stackdriver-sandbox.dev webpage off of github master branch
|
priority: p2 type: process
|
Currently stackdriver-sandbox.dev points to the content of the docs folder in the master branch. Whenever we make a change there, it's reflected on the website immediately. This prevents us from making incremental improvements that should not be visible to external customers until the release happens.
We should address that by pointing stackdriver-sandbox.dev to a hosted version of the website on AppEngine or somewhere else on GCP.
|
1.0
|
Move stackdriver-sandbox.dev webpage off of github master branch - Currently stackdriver-sandbox.dev points to the content of the docs folder in the master branch. Whenever we make a change there, it's reflected on the website immediately. This prevents us from making incremental improvements that should not be visible to external customers until the release happens.
We should address that by pointing stackdriver-sandbox.dev to a hosted version of the website on AppEngine or somewhere else on GCP.
|
process
|
move stackdriver sandbox dev webpage off of github master branch currently stackdriver sandbox dev points to the content of the docs folder in the master branch whenever we make a change there it s reflected on the website immediately this prevents us from making incremental improvements that should not be visible to external customers until the release happens we should address that by pointing stackdriver sandbox dev to a hosted version of the website on appengine or somewhere else on gcp
| 1
|
5,634
| 8,485,285,383
|
IssuesEvent
|
2018-10-26 07:14:53
|
thewca/wca-regulations
|
https://api.github.com/repos/thewca/wca-regulations
|
opened
|
Temporary "100-week" events
|
announcement-worthy process proposal
|
I would like to seriously consider [Arnaud's suggestion of a temporary event](https://www.speedsolving.com/forum/threads/idea-alternating-1-year-event.70659/). He suggests 1 year, but I think we should go for ≈2 years so that every puzzle can have a world champion once. 2 years minus a month of gap sounds appropriate, so I'm calling these "100-week events" in my head. :-D
We have had several discussions that touch on the issue that that adding or removing events is a big deal. Major competitions have a lot of pressure to hold all events, and the combination of 1) modern scheduling and 2) the lack of a maintained database for unofficial events at official competitions means that unofficial events are very uncommon.
Rough idea for an event that would be active from 2020 to 2021:
- In January 2019 we announce some number (say, ≈3) of events that we want to consider for the next cycle. We encourage competitions to hold these unofficially in order to work out the format and give the competitors a chance to try it out.
- Around August (after the major competitions are held), we hold a vote of all active WCA competitors for one event.
- The event becomes an official event from January 2020 to the end of 2021.
- In January 2021, we start the cycle for a new puzzle.
We have already retired events in the past, so this shouldn't require special tweaking from the website. What I really like about this idea is that it lowers the commitment for adding an official event, without setting unfair expectations and fueling debates about which events are worth keeping. It also prevents keeping around an event for so long that
I believe that new events are a benefit to the community because:
- We regularly get questions or requests for new events, so competitors want them! Listening to community wishes may become more important as we try to stay relevant compared to other efforts like Red Bull.
- If we regularly engage with the community for this one point of feedback, we have a chance to keep community conversations going between WCA officials and competitors. (I feel like we have a tendency for the two to disconnect over time.)
- The cubing community is large enough that puzzle manufacturers are likely to invest resources into puzzles that we consider for 100-week events. And once a well-engineered puzzle is on the market, it will hopefully stay available for longer. I would love for us to incentivize perpetual innovation in new types of speed puzzles.
- Novel puzzles encourage research into algorithms and software development, whose effects also stay around. e now see a lot of cubers focused on advanced techniques and tweaks to existing methods, but I think we're losing out on more classical development of methods.
- I would really like to see us explore less classic event formats, such as team BLD or FMC duel.
|
1.0
|
Temporary "100-week" events - I would like to seriously consider [Arnaud's suggestion of a temporary event](https://www.speedsolving.com/forum/threads/idea-alternating-1-year-event.70659/). He suggests 1 year, but I think we should go for ≈2 years so that every puzzle can have a world champion once. 2 years minus a month of gap sounds appropriate, so I'm calling these "100-week events" in my head. :-D
We have had several discussions that touch on the issue that that adding or removing events is a big deal. Major competitions have a lot of pressure to hold all events, and the combination of 1) modern scheduling and 2) the lack of a maintained database for unofficial events at official competitions means that unofficial events are very uncommon.
Rough idea for an event that would be active from 2020 to 2021:
- In January 2019 we announce some number (say, ≈3) of events that we want to consider for the next cycle. We encourage competitions to hold these unofficially in order to work out the format and give the competitors a chance to try it out.
- Around August (after the major competitions are held), we hold a vote of all active WCA competitors for one event.
- The event becomes an official event from January 2020 to the end of 2021.
- In January 2021, we start the cycle for a new puzzle.
We have already retired events in the past, so this shouldn't require special tweaking from the website. What I really like about this idea is that it lowers the commitment for adding an official event, without setting unfair expectations and fueling debates about which events are worth keeping. It also prevents keeping around an event for so long that
I believe that new events are a benefit to the community because:
- We regularly get questions or requests for new events, so competitors want them! Listening to community wishes may become more important as we try to stay relevant compared to other efforts like Red Bull.
- If we regularly engage with the community for this one point of feedback, we have a chance to keep community conversations going between WCA officials and competitors. (I feel like we have a tendency for the two to disconnect over time.)
- The cubing community is large enough that puzzle manufacturers are likely to invest resources into puzzles that we consider for 100-week events. And once a well-engineered puzzle is on the market, it will hopefully stay available for longer. I would love for us to incentivize perpetual innovation in new types of speed puzzles.
- Novel puzzles encourage research into algorithms and software development, whose effects also stay around. e now see a lot of cubers focused on advanced techniques and tweaks to existing methods, but I think we're losing out on more classical development of methods.
- I would really like to see us explore less classic event formats, such as team BLD or FMC duel.
|
process
|
temporary week events i would like to seriously consider he suggests year but i think we should go for ≈ years so that every puzzle can have a world champion once years minus a month of gap sounds appropriate so i m calling these week events in my head d we have had several discussions that touch on the issue that that adding or removing events is a big deal major competitions have a lot of pressure to hold all events and the combination of modern scheduling and the lack of a maintained database for unofficial events at official competitions means that unofficial events are very uncommon rough idea for an event that would be active from to in january we announce some number say ≈ of events that we want to consider for the next cycle we encourage competitions to hold these unofficially in order to work out the format and give the competitors a chance to try it out around august after the major competitions are held we hold a vote of all active wca competitors for one event the event becomes an official event from january to the end of in january we start the cycle for a new puzzle we have already retired events in the past so this shouldn t require special tweaking from the website what i really like about this idea is that it lowers the commitment for adding an official event without setting unfair expectations and fueling debates about which events are worth keeping it also prevents keeping around an event for so long that i believe that new events are a benefit to the community because we regularly get questions or requests for new events so competitors want them listening to community wishes may become more important as we try to stay relevant compared to other efforts like red bull if we regularly engage with the community for this one point of feedback we have a chance to keep community conversations going between wca officials and competitors i feel like we have a tendency for the two to disconnect over time the cubing community is large enough that puzzle manufacturers are likely to invest resources into puzzles that we consider for week events and once a well engineered puzzle is on the market it will hopefully stay available for longer i would love for us to incentivize perpetual innovation in new types of speed puzzles novel puzzles encourage research into algorithms and software development whose effects also stay around e now see a lot of cubers focused on advanced techniques and tweaks to existing methods but i think we re losing out on more classical development of methods i would really like to see us explore less classic event formats such as team bld or fmc duel
| 1
|
10,169
| 8,403,543,181
|
IssuesEvent
|
2018-10-11 10:03:42
|
nest/nest-simulator
|
https://api.github.com/repos/nest/nest-simulator
|
closed
|
Index of user contributed modules/extensions (MyModule)
|
C: Infrastructure I: No breaking change P: In progess S: Low T: Maintenance
|
As far as I can see it there is currently no index for user contributed modules hosted on the nest website or the github wiki (https://github.com/nest/nest-simulator/wiki).
As more specialized modules are contributed I think it would make sense to maintain a list (possibly sorted by use-case) to point the community to available modules.
|
1.0
|
Index of user contributed modules/extensions (MyModule) - As far as I can see it there is currently no index for user contributed modules hosted on the nest website or the github wiki (https://github.com/nest/nest-simulator/wiki).
As more specialized modules are contributed I think it would make sense to maintain a list (possibly sorted by use-case) to point the community to available modules.
|
non_process
|
index of user contributed modules extensions mymodule as far as i can see it there is currently no index for user contributed modules hosted on the nest website or the github wiki as more specialized modules are contributed i think it would make sense to maintain a list possibly sorted by use case to point the community to available modules
| 0
|
5,268
| 8,058,213,508
|
IssuesEvent
|
2018-08-02 17:45:41
|
dita-ot/dita-ot
|
https://api.github.com/repos/dita-ot/dita-ot
|
closed
|
[DOTX031E][ERROR] The file 'topic.dita' is not available to resolve link information in XHTML transformation when using chunk='to-content' feature
|
obsolete preprocess/chunking
|
The XHTML transformation fails when a DITA map uses the `chunk='to-content'` feature and there are related-links to the merged topics. The error that I get is:
```
[topicpull] Recoverable error on line 576 of topicpullImpl.xsl:
[topicpull] FODC0002: java.io.FileNotFoundException:
[topicpull] C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml\topics\flowers\concepts\springFlowers.dita (The system cannot find the path specified)
[topicpull] file:/C:/Users/radu_pisoi/Desktop/temp/dita-issue/issue2/flowers/topics/flowers/snowdrop.dita:9:33: [DOTX031E][ERROR]: The file file:/C:/Users/radu_pisoi/Desktop/temp/dita-issue/issue2/flowers/temp/xhtml/topics/flowers/concepts/springFlowers.dita is not available to resolve link information.
```
I noticed that if I remove the related-link to the topic merged in the parent topic, then the issue disappears.
```
Executing:
"D:\workspace\eXml\jre\jre1.8.0_72-x64/bin/java" "-Dhttp.proxySet=true" "-Dhttp.proxyHost=proxy.sync.ro" "-Dhttps.proxyHost=proxy.sync.ro" "-Dhttp.proxyPort=3128" "-Dhttps.proxyPort=3128" -Xmx384m "-Doxygen.org.apache.xerces.xni.parser.XMLParserConfiguration=org.ditang.relaxng.defaults.RelaxDefaultsParserConfiguration" -classpath "D:\workspace\eXml\tools\ant/lib/ant-launcher.jar" "-Dant.home=D:\workspace\eXml\tools\ant" org.apache.tools.ant.launch.Launcher -lib "D:\workspace\dita-ot-2.3.1\lib\saxon-9.1.0.8-dom.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\saxon-9.1.0.8.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\commons-io-2.4.jar" -lib "D:\workspace\eXml\classes" -lib "D:\workspace\eXml\lib\oxygen-content-completion-api.jar" -lib "D:\workspace\eXml\lib\oxygen-css-pretty-printer.jar" -lib "D:\workspace\eXml\lib\oxygen-css-validator.jar" -lib "D:\workspace\eXml\lib\oxygen-emf.jar" -lib "D:\workspace\eXml\lib\oxygen-jfx-components.jar" -lib "D:\workspace\eXml\lib\oxygen-text-search.jar" -lib "D:\workspace\eXml\lib\oxygen-token-markers.jar" -lib "D:\workspace\eXml\lib\oxygen-validation-api.jar" -lib "D:\workspace\eXml\lib\oxygen-xquery-pretty-printer.jar" -lib "D:\workspace\eXml\lib\resolver.jar" -lib "D:\workspace\eXml\lib\oxygen-token-markers.jar" -lib "D:\workspace\eXml\lib\org.eclipse.wst.xml.xpath2.processor_1.2.0.jar" -lib "D:\workspace\eXml\lib\xml-apis.jar" -lib "D:\workspace\eXml\lib\xercesImpl.jar" -lib "D:\workspace\eXml\lib\commons-io-1.3.1.jar" -lib "D:\workspace\eXml\lib\commons-logging-1.2.jar" -lib "D:\workspace\eXml\lib\log4j.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\commons-codec-1.9.jar" -lib "D:\workspace\eXml\lib\jing.jar" -lib "D:\workspace\eXml\lib\saxon9ee.jar" -lib "D:\workspace\eXml\lib\saxon.jar" -lib "D:\workspace\eXml\lib\xmlgraphics-commons-2.1.jar" -lib "D:\workspace\eXml\lib\fop.jar" -lib "D:\workspace\eXml\lib\batik-all-1.8.jar" -lib "D:\workspace\eXml\lib\js.jar" -lib "D:\workspace\eXml\lib\poi-3.10-FINAL-20140208.jar" -lib "D:\workspace\eXml\lib\nekohtml.jar" -lib "D:\workspace\eXml\lib\xml-apis-ext.jar" -lib "D:\workspace\eXml\lib\avalon-framework-api-4.3.1.jar" -lib "D:\workspace\eXml\lib\avalon-framework-impl-4.3.1.jar" -lib "D:\workspace\eXml\lib\jeuclid-core.jar" -lib "D:\workspace\eXml\lib\jeuclid-fop.jar" -lib "D:\workspace\eXml\lib\jai_tiff.jar" -lib "D:\workspace\eXml\lib\jh.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\jsearch.jar" -lib "D:\workspace\eXml\lib\lucene-analyzers-common-6.1.0.jar" -lib "D:\workspace\eXml\lib\lucene-analyzers-kuromoji-6.1.0.jar" -lib "D:\workspace\eXml\lib\lucene-core-6.1.0.jar" -lib "D:\workspace\eXml\lib\lucene-misc-6.1.0.jar" -lib "D:\workspace\eXml\lib\lucene-queries-6.1.0.jar" -lib "D:\workspace\eXml\lib\lucene-queryparser-6.1.0.jar" -lib "D:\workspace\eXml\lib\lucene-suggest-6.1.0.jar" -lib "D:\workspace\dita-ot-2.3.1" -lib "D:\workspace\dita-ot-2.3.1\lib\dost-patches.jar" -lib "D:\workspace\dita-ot-2.3.1\lib" -lib "D:\workspace\dita-ot-2.3.1\lib\ant-apache-resolver-1.9.4.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\ant-launcher.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\ant.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\commons-codec-1.9.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\commons-io-2.4.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\dost-configuration.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\dost.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\guava-19.0.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\icu4j-54.1.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\saxon-9.1.0.8-dom.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\saxon-9.1.0.8.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\xercesImpl-2.11.0.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\xml-apis-1.4.01.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\xml-resolver-1.2.jar" -lib "D:\workspace\dita-ot-2.3.1\plugins\org.dita.odt\lib\odt.jar" -lib "D:\workspace\dita-ot-2.3.1\plugins\org.dita.pdf2\lib\fo.jar" -lib "D:\workspace\dita-ot-2.3.1\plugins\org.dita.pdf2.axf\lib\axf.jar" -lib "D:\workspace\dita-ot-2.3.1\plugins\org.dita.pdf2.xep\lib\xep.jar" -lib "D:\workspace\dita-ot-2.3.1\plugins\org.dita.wordrtf\lib\wordrtf.jar" -f "D:\workspace\dita-ot-2.3.1/build.xml" "-Dtranstype=xhtml" "-Dbasedir=C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers" "-Doutput.dir=C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers/out/xhtml" "-Ddita.temp.dir=C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers/temp/xhtml" "-Dargs.input=C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\flowers.ditamap" "-Ddita.dir=D:\workspace\dita-ot-2.3.1" "-DbaseJVMArgLine=-Xmx384m"
Buildfile: D:\workspace\dita-ot-2.3.1\build.xml
init:
dita2xhtml.init:
check-arg:
[mkdir] Created dir: C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\out\xhtml
[mkdir] Created dir: C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml
log-arg:
[echo] *****************************************************************
[echo] * basedir = C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers
[echo] * dita.dir = D:\workspace\dita-ot-2.3.1
[echo] * transtype = xhtml
[echo] * tempdir = C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers/temp/xhtml
[echo] * outputdir = C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers/out/xhtml
[echo] * clean.temp = true
[echo] * DITA-OT version = 2.3.1
[echo] * XML parser = Xerces
[echo] * XSLT processor = Saxon
[echo] * collator = ICU
[echo] *****************************************************************
[echo] #Ant properties
[echo] #Fri Aug 19 17:37:23 EEST 2016
[echo] args.grammar.cache=yes
[echo] args.input=C\:\\Users\\radu_pisoi\\Desktop\\temp\\dita-issue\\issue2\\flowers\\flowers.ditamap
[echo] args.logdir=C\:\\Users\\radu_pisoi\\Desktop\\temp\\dita-issue\\issue2\\flowers/out/xhtml
[echo] args.xml.systemid.set=yes
[echo] dita.dir=D\:\\workspace\\dita-ot-2.3.1
[echo] dita.plugin.com.sophos.tocjs.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\com.sophos.tocjs
[echo] dita.plugin.org.dita.base.dir=D\:\\workspace\\dita-ot-2.3.1
[echo] dita.plugin.org.dita.docbook.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.docbook
[echo] dita.plugin.org.dita.eclipsecontent.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.eclipsecontent
[echo] dita.plugin.org.dita.eclipsehelp.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.eclipsehelp
[echo] dita.plugin.org.dita.html5.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.html5
[echo] dita.plugin.org.dita.htmlhelp.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.htmlhelp
[echo] dita.plugin.org.dita.javahelp.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.javahelp
[echo] dita.plugin.org.dita.odt.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.odt
[echo] dita.plugin.org.dita.pdf2.axf.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.pdf2.axf
[echo] dita.plugin.org.dita.pdf2.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.pdf2
[echo] dita.plugin.org.dita.pdf2.fop.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.pdf2.fop
[echo] dita.plugin.org.dita.pdf2.xep.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.pdf2.xep
[echo] dita.plugin.org.dita.specialization.dita11.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.specialization.dita11
[echo] dita.plugin.org.dita.specialization.eclipsemap.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.specialization.eclipsemap
[echo] dita.plugin.org.dita.troff.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.troff
[echo] dita.plugin.org.dita.wordrtf.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.wordrtf
[echo] dita.plugin.org.dita.xhtml.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.xhtml
[echo] dita.plugin.org.oasis-open.dita.v1_2.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.oasis-open.dita.v1_2
[echo] dita.plugin.org.oasis-open.dita.v1_3.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.oasis-open.dita.v1_3
[echo] dita.temp.dir=C\:\\Users\\radu_pisoi\\Desktop\\temp\\dita-issue\\issue2\\flowers/temp/xhtml
[echo] *****************************************************************
--------------------------------------------------
topicpull-check:
topicpull:
[topicpull] Transforming into C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml
[topicpull] Loading stylesheet D:\workspace\dita-ot-2.3.1\xsl\preprocess\topicpull.xsl
[topicpull] Processing C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml\topics\index.dita
[topicpull] Processing C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml\concepts\springFlowers.dita
[topicpull] Recoverable error on line 576 of topicpullImpl.xsl:
[topicpull] FODC0002: java.io.FileNotFoundException:
[topicpull] C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml\topics\flowers\concepts\springFlowers.dita (The system cannot find the path specified)
[topicpull] file:/C:/Users/radu_pisoi/Desktop/temp/dita-issue/issue2/flowers/topics/flowers/snowdrop.dita:9:33: [DOTX031E][ERROR]: The file file:/C:/Users/radu_pisoi/Desktop/temp/dita-issue/issue2/flowers/temp/xhtml/topics/flowers/concepts/springFlowers.dita is not available to resolve link information.
[topicpull] Recoverable error on line 654 of topicpullImpl.xsl:
[topicpull] FODC0002: java.io.FileNotFoundException:
[topicpull] C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml\topics\flowers\concepts\springFlowers.dita (The system cannot find the path specified)
[topicpull] Recoverable error on line 1108 of topicpullImpl.xsl:
[topicpull] FODC0002: java.io.FileNotFoundException:
[topicpull] C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml\topics\flowers\concepts\springFlowers.dita (The system cannot find the path specified)
[topicpull] Recoverable error on line 1844 of topicpullImpl.xsl:
[topicpull] FODC0002: java.io.FileNotFoundException:
[topicpull] C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml\topics\flowers\concepts\springFlowers.dita (The system cannot find the path specified)
```
I've tested with DITA-OT 2.3.1 and the attached DITA map:
[flowers.zip](https://github.com/dita-ot/dita-ot/files/427257/flowers.zip)
|
1.0
|
[DOTX031E][ERROR] The file 'topic.dita' is not available to resolve link information in XHTML transformation when using chunk='to-content' feature - The XHTML transformation fails when a DITA map uses the `chunk='to-content'` feature and there are related-links to the merged topics. The error that I get is:
```
[topicpull] Recoverable error on line 576 of topicpullImpl.xsl:
[topicpull] FODC0002: java.io.FileNotFoundException:
[topicpull] C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml\topics\flowers\concepts\springFlowers.dita (The system cannot find the path specified)
[topicpull] file:/C:/Users/radu_pisoi/Desktop/temp/dita-issue/issue2/flowers/topics/flowers/snowdrop.dita:9:33: [DOTX031E][ERROR]: The file file:/C:/Users/radu_pisoi/Desktop/temp/dita-issue/issue2/flowers/temp/xhtml/topics/flowers/concepts/springFlowers.dita is not available to resolve link information.
```
I noticed that if I remove the related-link to the topic merged in the parent topic, then the issue disappears.
```
Executing:
"D:\workspace\eXml\jre\jre1.8.0_72-x64/bin/java" "-Dhttp.proxySet=true" "-Dhttp.proxyHost=proxy.sync.ro" "-Dhttps.proxyHost=proxy.sync.ro" "-Dhttp.proxyPort=3128" "-Dhttps.proxyPort=3128" -Xmx384m "-Doxygen.org.apache.xerces.xni.parser.XMLParserConfiguration=org.ditang.relaxng.defaults.RelaxDefaultsParserConfiguration" -classpath "D:\workspace\eXml\tools\ant/lib/ant-launcher.jar" "-Dant.home=D:\workspace\eXml\tools\ant" org.apache.tools.ant.launch.Launcher -lib "D:\workspace\dita-ot-2.3.1\lib\saxon-9.1.0.8-dom.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\saxon-9.1.0.8.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\commons-io-2.4.jar" -lib "D:\workspace\eXml\classes" -lib "D:\workspace\eXml\lib\oxygen-content-completion-api.jar" -lib "D:\workspace\eXml\lib\oxygen-css-pretty-printer.jar" -lib "D:\workspace\eXml\lib\oxygen-css-validator.jar" -lib "D:\workspace\eXml\lib\oxygen-emf.jar" -lib "D:\workspace\eXml\lib\oxygen-jfx-components.jar" -lib "D:\workspace\eXml\lib\oxygen-text-search.jar" -lib "D:\workspace\eXml\lib\oxygen-token-markers.jar" -lib "D:\workspace\eXml\lib\oxygen-validation-api.jar" -lib "D:\workspace\eXml\lib\oxygen-xquery-pretty-printer.jar" -lib "D:\workspace\eXml\lib\resolver.jar" -lib "D:\workspace\eXml\lib\oxygen-token-markers.jar" -lib "D:\workspace\eXml\lib\org.eclipse.wst.xml.xpath2.processor_1.2.0.jar" -lib "D:\workspace\eXml\lib\xml-apis.jar" -lib "D:\workspace\eXml\lib\xercesImpl.jar" -lib "D:\workspace\eXml\lib\commons-io-1.3.1.jar" -lib "D:\workspace\eXml\lib\commons-logging-1.2.jar" -lib "D:\workspace\eXml\lib\log4j.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\commons-codec-1.9.jar" -lib "D:\workspace\eXml\lib\jing.jar" -lib "D:\workspace\eXml\lib\saxon9ee.jar" -lib "D:\workspace\eXml\lib\saxon.jar" -lib "D:\workspace\eXml\lib\xmlgraphics-commons-2.1.jar" -lib "D:\workspace\eXml\lib\fop.jar" -lib "D:\workspace\eXml\lib\batik-all-1.8.jar" -lib "D:\workspace\eXml\lib\js.jar" -lib "D:\workspace\eXml\lib\poi-3.10-FINAL-20140208.jar" -lib "D:\workspace\eXml\lib\nekohtml.jar" -lib "D:\workspace\eXml\lib\xml-apis-ext.jar" -lib "D:\workspace\eXml\lib\avalon-framework-api-4.3.1.jar" -lib "D:\workspace\eXml\lib\avalon-framework-impl-4.3.1.jar" -lib "D:\workspace\eXml\lib\jeuclid-core.jar" -lib "D:\workspace\eXml\lib\jeuclid-fop.jar" -lib "D:\workspace\eXml\lib\jai_tiff.jar" -lib "D:\workspace\eXml\lib\jh.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\jsearch.jar" -lib "D:\workspace\eXml\lib\lucene-analyzers-common-6.1.0.jar" -lib "D:\workspace\eXml\lib\lucene-analyzers-kuromoji-6.1.0.jar" -lib "D:\workspace\eXml\lib\lucene-core-6.1.0.jar" -lib "D:\workspace\eXml\lib\lucene-misc-6.1.0.jar" -lib "D:\workspace\eXml\lib\lucene-queries-6.1.0.jar" -lib "D:\workspace\eXml\lib\lucene-queryparser-6.1.0.jar" -lib "D:\workspace\eXml\lib\lucene-suggest-6.1.0.jar" -lib "D:\workspace\dita-ot-2.3.1" -lib "D:\workspace\dita-ot-2.3.1\lib\dost-patches.jar" -lib "D:\workspace\dita-ot-2.3.1\lib" -lib "D:\workspace\dita-ot-2.3.1\lib\ant-apache-resolver-1.9.4.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\ant-launcher.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\ant.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\commons-codec-1.9.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\commons-io-2.4.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\dost-configuration.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\dost.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\guava-19.0.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\icu4j-54.1.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\saxon-9.1.0.8-dom.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\saxon-9.1.0.8.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\xercesImpl-2.11.0.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\xml-apis-1.4.01.jar" -lib "D:\workspace\dita-ot-2.3.1\lib\xml-resolver-1.2.jar" -lib "D:\workspace\dita-ot-2.3.1\plugins\org.dita.odt\lib\odt.jar" -lib "D:\workspace\dita-ot-2.3.1\plugins\org.dita.pdf2\lib\fo.jar" -lib "D:\workspace\dita-ot-2.3.1\plugins\org.dita.pdf2.axf\lib\axf.jar" -lib "D:\workspace\dita-ot-2.3.1\plugins\org.dita.pdf2.xep\lib\xep.jar" -lib "D:\workspace\dita-ot-2.3.1\plugins\org.dita.wordrtf\lib\wordrtf.jar" -f "D:\workspace\dita-ot-2.3.1/build.xml" "-Dtranstype=xhtml" "-Dbasedir=C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers" "-Doutput.dir=C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers/out/xhtml" "-Ddita.temp.dir=C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers/temp/xhtml" "-Dargs.input=C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\flowers.ditamap" "-Ddita.dir=D:\workspace\dita-ot-2.3.1" "-DbaseJVMArgLine=-Xmx384m"
Buildfile: D:\workspace\dita-ot-2.3.1\build.xml
init:
dita2xhtml.init:
check-arg:
[mkdir] Created dir: C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\out\xhtml
[mkdir] Created dir: C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml
log-arg:
[echo] *****************************************************************
[echo] * basedir = C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers
[echo] * dita.dir = D:\workspace\dita-ot-2.3.1
[echo] * transtype = xhtml
[echo] * tempdir = C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers/temp/xhtml
[echo] * outputdir = C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers/out/xhtml
[echo] * clean.temp = true
[echo] * DITA-OT version = 2.3.1
[echo] * XML parser = Xerces
[echo] * XSLT processor = Saxon
[echo] * collator = ICU
[echo] *****************************************************************
[echo] #Ant properties
[echo] #Fri Aug 19 17:37:23 EEST 2016
[echo] args.grammar.cache=yes
[echo] args.input=C\:\\Users\\radu_pisoi\\Desktop\\temp\\dita-issue\\issue2\\flowers\\flowers.ditamap
[echo] args.logdir=C\:\\Users\\radu_pisoi\\Desktop\\temp\\dita-issue\\issue2\\flowers/out/xhtml
[echo] args.xml.systemid.set=yes
[echo] dita.dir=D\:\\workspace\\dita-ot-2.3.1
[echo] dita.plugin.com.sophos.tocjs.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\com.sophos.tocjs
[echo] dita.plugin.org.dita.base.dir=D\:\\workspace\\dita-ot-2.3.1
[echo] dita.plugin.org.dita.docbook.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.docbook
[echo] dita.plugin.org.dita.eclipsecontent.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.eclipsecontent
[echo] dita.plugin.org.dita.eclipsehelp.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.eclipsehelp
[echo] dita.plugin.org.dita.html5.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.html5
[echo] dita.plugin.org.dita.htmlhelp.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.htmlhelp
[echo] dita.plugin.org.dita.javahelp.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.javahelp
[echo] dita.plugin.org.dita.odt.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.odt
[echo] dita.plugin.org.dita.pdf2.axf.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.pdf2.axf
[echo] dita.plugin.org.dita.pdf2.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.pdf2
[echo] dita.plugin.org.dita.pdf2.fop.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.pdf2.fop
[echo] dita.plugin.org.dita.pdf2.xep.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.pdf2.xep
[echo] dita.plugin.org.dita.specialization.dita11.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.specialization.dita11
[echo] dita.plugin.org.dita.specialization.eclipsemap.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.specialization.eclipsemap
[echo] dita.plugin.org.dita.troff.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.troff
[echo] dita.plugin.org.dita.wordrtf.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.wordrtf
[echo] dita.plugin.org.dita.xhtml.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.dita.xhtml
[echo] dita.plugin.org.oasis-open.dita.v1_2.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.oasis-open.dita.v1_2
[echo] dita.plugin.org.oasis-open.dita.v1_3.dir=D\:\\workspace\\dita-ot-2.3.1\\plugins\\org.oasis-open.dita.v1_3
[echo] dita.temp.dir=C\:\\Users\\radu_pisoi\\Desktop\\temp\\dita-issue\\issue2\\flowers/temp/xhtml
[echo] *****************************************************************
--------------------------------------------------
topicpull-check:
topicpull:
[topicpull] Transforming into C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml
[topicpull] Loading stylesheet D:\workspace\dita-ot-2.3.1\xsl\preprocess\topicpull.xsl
[topicpull] Processing C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml\topics\index.dita
[topicpull] Processing C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml\concepts\springFlowers.dita
[topicpull] Recoverable error on line 576 of topicpullImpl.xsl:
[topicpull] FODC0002: java.io.FileNotFoundException:
[topicpull] C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml\topics\flowers\concepts\springFlowers.dita (The system cannot find the path specified)
[topicpull] file:/C:/Users/radu_pisoi/Desktop/temp/dita-issue/issue2/flowers/topics/flowers/snowdrop.dita:9:33: [DOTX031E][ERROR]: The file file:/C:/Users/radu_pisoi/Desktop/temp/dita-issue/issue2/flowers/temp/xhtml/topics/flowers/concepts/springFlowers.dita is not available to resolve link information.
[topicpull] Recoverable error on line 654 of topicpullImpl.xsl:
[topicpull] FODC0002: java.io.FileNotFoundException:
[topicpull] C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml\topics\flowers\concepts\springFlowers.dita (The system cannot find the path specified)
[topicpull] Recoverable error on line 1108 of topicpullImpl.xsl:
[topicpull] FODC0002: java.io.FileNotFoundException:
[topicpull] C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml\topics\flowers\concepts\springFlowers.dita (The system cannot find the path specified)
[topicpull] Recoverable error on line 1844 of topicpullImpl.xsl:
[topicpull] FODC0002: java.io.FileNotFoundException:
[topicpull] C:\Users\radu_pisoi\Desktop\temp\dita-issue\issue2\flowers\temp\xhtml\topics\flowers\concepts\springFlowers.dita (The system cannot find the path specified)
```
I've tested with DITA-OT 2.3.1 and the attached DITA map:
[flowers.zip](https://github.com/dita-ot/dita-ot/files/427257/flowers.zip)
|
process
|
the file topic dita is not available to resolve link information in xhtml transformation when using chunk to content feature the xhtml transformation fails when a dita map uses the chunk to content feature and there are related links to the merged topics the error that i get is recoverable error on line of topicpullimpl xsl java io filenotfoundexception c users radu pisoi desktop temp dita issue flowers temp xhtml topics flowers concepts springflowers dita the system cannot find the path specified file c users radu pisoi desktop temp dita issue flowers topics flowers snowdrop dita the file file c users radu pisoi desktop temp dita issue flowers temp xhtml topics flowers concepts springflowers dita is not available to resolve link information i noticed that if i remove the related link to the topic merged in the parent topic then the issue disappears executing d workspace exml jre bin java dhttp proxyset true dhttp proxyhost proxy sync ro dhttps proxyhost proxy sync ro dhttp proxyport dhttps proxyport doxygen org apache xerces xni parser xmlparserconfiguration org ditang relaxng defaults relaxdefaultsparserconfiguration classpath d workspace exml tools ant lib ant launcher jar dant home d workspace exml tools ant org apache tools ant launch launcher lib d workspace dita ot lib saxon dom jar lib d workspace dita ot lib saxon jar lib d workspace dita ot lib commons io jar lib d workspace exml classes lib d workspace exml lib oxygen content completion api jar lib d workspace exml lib oxygen css pretty printer jar lib d workspace exml lib oxygen css validator jar lib d workspace exml lib oxygen emf jar lib d workspace exml lib oxygen jfx components jar lib d workspace exml lib oxygen text search jar lib d workspace exml lib oxygen token markers jar lib d workspace exml lib oxygen validation api jar lib d workspace exml lib oxygen xquery pretty printer jar lib d workspace exml lib resolver jar lib d workspace exml lib oxygen token markers jar lib d workspace exml lib org eclipse wst xml processor jar lib d workspace exml lib xml apis jar lib d workspace exml lib xercesimpl jar lib d workspace exml lib commons io jar lib d workspace exml lib commons logging jar lib d workspace exml lib jar lib d workspace dita ot lib commons codec jar lib d workspace exml lib jing jar lib d workspace exml lib jar lib d workspace exml lib saxon jar lib d workspace exml lib xmlgraphics commons jar lib d workspace exml lib fop jar lib d workspace exml lib batik all jar lib d workspace exml lib js jar lib d workspace exml lib poi final jar lib d workspace exml lib nekohtml jar lib d workspace exml lib xml apis ext jar lib d workspace exml lib avalon framework api jar lib d workspace exml lib avalon framework impl jar lib d workspace exml lib jeuclid core jar lib d workspace exml lib jeuclid fop jar lib d workspace exml lib jai tiff jar lib d workspace exml lib jh jar lib d workspace dita ot lib jsearch jar lib d workspace exml lib lucene analyzers common jar lib d workspace exml lib lucene analyzers kuromoji jar lib d workspace exml lib lucene core jar lib d workspace exml lib lucene misc jar lib d workspace exml lib lucene queries jar lib d workspace exml lib lucene queryparser jar lib d workspace exml lib lucene suggest jar lib d workspace dita ot lib d workspace dita ot lib dost patches jar lib d workspace dita ot lib lib d workspace dita ot lib ant apache resolver jar lib d workspace dita ot lib ant launcher jar lib d workspace dita ot lib ant jar lib d workspace dita ot lib commons codec jar lib d workspace dita ot lib commons io jar lib d workspace dita ot lib dost configuration jar lib d workspace dita ot lib dost jar lib d workspace dita ot lib guava jar lib d workspace dita ot lib jar lib d workspace dita ot lib saxon dom jar lib d workspace dita ot lib saxon jar lib d workspace dita ot lib xercesimpl jar lib d workspace dita ot lib xml apis jar lib d workspace dita ot lib xml resolver jar lib d workspace dita ot plugins org dita odt lib odt jar lib d workspace dita ot plugins org dita lib fo jar lib d workspace dita ot plugins org dita axf lib axf jar lib d workspace dita ot plugins org dita xep lib xep jar lib d workspace dita ot plugins org dita wordrtf lib wordrtf jar f d workspace dita ot build xml dtranstype xhtml dbasedir c users radu pisoi desktop temp dita issue flowers doutput dir c users radu pisoi desktop temp dita issue flowers out xhtml ddita temp dir c users radu pisoi desktop temp dita issue flowers temp xhtml dargs input c users radu pisoi desktop temp dita issue flowers flowers ditamap ddita dir d workspace dita ot dbasejvmargline buildfile d workspace dita ot build xml init init check arg created dir c users radu pisoi desktop temp dita issue flowers out xhtml created dir c users radu pisoi desktop temp dita issue flowers temp xhtml log arg basedir c users radu pisoi desktop temp dita issue flowers dita dir d workspace dita ot transtype xhtml tempdir c users radu pisoi desktop temp dita issue flowers temp xhtml outputdir c users radu pisoi desktop temp dita issue flowers out xhtml clean temp true dita ot version xml parser xerces xslt processor saxon collator icu ant properties fri aug eest args grammar cache yes args input c users radu pisoi desktop temp dita issue flowers flowers ditamap args logdir c users radu pisoi desktop temp dita issue flowers out xhtml args xml systemid set yes dita dir d workspace dita ot dita plugin com sophos tocjs dir d workspace dita ot plugins com sophos tocjs dita plugin org dita base dir d workspace dita ot dita plugin org dita docbook dir d workspace dita ot plugins org dita docbook dita plugin org dita eclipsecontent dir d workspace dita ot plugins org dita eclipsecontent dita plugin org dita eclipsehelp dir d workspace dita ot plugins org dita eclipsehelp dita plugin org dita dir d workspace dita ot plugins org dita dita plugin org dita htmlhelp dir d workspace dita ot plugins org dita htmlhelp dita plugin org dita javahelp dir d workspace dita ot plugins org dita javahelp dita plugin org dita odt dir d workspace dita ot plugins org dita odt dita plugin org dita axf dir d workspace dita ot plugins org dita axf dita plugin org dita dir d workspace dita ot plugins org dita dita plugin org dita fop dir d workspace dita ot plugins org dita fop dita plugin org dita xep dir d workspace dita ot plugins org dita xep dita plugin org dita specialization dir d workspace dita ot plugins org dita specialization dita plugin org dita specialization eclipsemap dir d workspace dita ot plugins org dita specialization eclipsemap dita plugin org dita troff dir d workspace dita ot plugins org dita troff dita plugin org dita wordrtf dir d workspace dita ot plugins org dita wordrtf dita plugin org dita xhtml dir d workspace dita ot plugins org dita xhtml dita plugin org oasis open dita dir d workspace dita ot plugins org oasis open dita dita plugin org oasis open dita dir d workspace dita ot plugins org oasis open dita dita temp dir c users radu pisoi desktop temp dita issue flowers temp xhtml topicpull check topicpull transforming into c users radu pisoi desktop temp dita issue flowers temp xhtml loading stylesheet d workspace dita ot xsl preprocess topicpull xsl processing c users radu pisoi desktop temp dita issue flowers temp xhtml topics index dita processing c users radu pisoi desktop temp dita issue flowers temp xhtml concepts springflowers dita recoverable error on line of topicpullimpl xsl java io filenotfoundexception c users radu pisoi desktop temp dita issue flowers temp xhtml topics flowers concepts springflowers dita the system cannot find the path specified file c users radu pisoi desktop temp dita issue flowers topics flowers snowdrop dita the file file c users radu pisoi desktop temp dita issue flowers temp xhtml topics flowers concepts springflowers dita is not available to resolve link information recoverable error on line of topicpullimpl xsl java io filenotfoundexception c users radu pisoi desktop temp dita issue flowers temp xhtml topics flowers concepts springflowers dita the system cannot find the path specified recoverable error on line of topicpullimpl xsl java io filenotfoundexception c users radu pisoi desktop temp dita issue flowers temp xhtml topics flowers concepts springflowers dita the system cannot find the path specified recoverable error on line of topicpullimpl xsl java io filenotfoundexception c users radu pisoi desktop temp dita issue flowers temp xhtml topics flowers concepts springflowers dita the system cannot find the path specified i ve tested with dita ot and the attached dita map
| 1
|
9,477
| 12,475,855,306
|
IssuesEvent
|
2020-05-29 12:26:31
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Possibility to create vector tiles directly from QGIS
|
Feature Request Processing Vector tiles
|
Ideally a processing algorithm (similar to mbtiles/xyz tiles) to generate vector tiles from QGIS map.
|
1.0
|
Possibility to create vector tiles directly from QGIS - Ideally a processing algorithm (similar to mbtiles/xyz tiles) to generate vector tiles from QGIS map.
|
process
|
possibility to create vector tiles directly from qgis ideally a processing algorithm similar to mbtiles xyz tiles to generate vector tiles from qgis map
| 1
|
15,242
| 19,179,995,635
|
IssuesEvent
|
2021-12-04 07:50:01
|
varabyte/kobweb
|
https://api.github.com/repos/varabyte/kobweb
|
closed
|
Audit ktor plugins
|
process
|
When I created the initial project, I selected a bunch of plugins that seemed relevant, but right now they may not be being used, and some of them are generating compiler warnings.
|
1.0
|
Audit ktor plugins - When I created the initial project, I selected a bunch of plugins that seemed relevant, but right now they may not be being used, and some of them are generating compiler warnings.
|
process
|
audit ktor plugins when i created the initial project i selected a bunch of plugins that seemed relevant but right now they may not be being used and some of them are generating compiler warnings
| 1
|
5,952
| 8,775,486,062
|
IssuesEvent
|
2018-12-18 23:19:48
|
knative/serving
|
https://api.github.com/repos/knative/serving
|
closed
|
Replicate knative-releases to {us,asia,eu}.gcr.io
|
area/test-and-release kind/process
|
<!--
Pro-tip: You can leave this block commented, and it still works!
Select the appropriate areas for your issue:
/area test-and-release
Classify what kind of issue this is:
/kind process
-->
## Expected Behavior
Our releases should be replicated to different regions to improve availability and spacial locality.
## Actual Behavior
Our releases are published only to gcr.io.
|
1.0
|
Replicate knative-releases to {us,asia,eu}.gcr.io - <!--
Pro-tip: You can leave this block commented, and it still works!
Select the appropriate areas for your issue:
/area test-and-release
Classify what kind of issue this is:
/kind process
-->
## Expected Behavior
Our releases should be replicated to different regions to improve availability and spacial locality.
## Actual Behavior
Our releases are published only to gcr.io.
|
process
|
replicate knative releases to us asia eu gcr io pro tip you can leave this block commented and it still works select the appropriate areas for your issue area test and release classify what kind of issue this is kind process expected behavior our releases should be replicated to different regions to improve availability and spacial locality actual behavior our releases are published only to gcr io
| 1
|
5,598
| 8,459,928,961
|
IssuesEvent
|
2018-10-22 17:19:29
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
Change specific responses to chemicals to part_of response to X
|
cellular processes
|
During the [June 2014 meeting on annotation extension relations](http://wiki.geneontology.org/index.php/Annotation_Extension_meeting_2014-06-16),
it was agreed to change terms for specific responses to chemicals (see
[item 1 from 140624 annotation call minutes](http://wiki.geneontology.org/index.php/Annotation_Conf._Call,_June_24,_2014#Summary_of_proposed_action_items_from_annotation_extension_relation_meeting)).
Examples:
[chemosensory behavior](http://purl.obolibrary.org/obo/GO_0007635)
is_a 'response to chemical' ->
[chemosensory behavior](http://purl.obolibrary.org/obo/GO_0007635)
part_of 'response to chemical'
Motivating example (seems to be fictional term):
‘proteolysis [involved] in cellular response to drug' could have two has_input relationships:
has_input: drug
has_input: proteolysis target.
The has_inputs work in the individual cases but when combined, how do you know which input is which?
A. The drug isn’t an input to the proteolysis. The proteolysis is part of the cellular response to drug. So 'proteolysis [involved] in cellular response to drug' should be part of cellular response to drug, rather than is_a.
Reported by: dosumis
Original Ticket: [geneontology/ontology-requests/10958](https://sourceforge.net/p/geneontology/ontology-requests/10958)
|
1.0
|
Change specific responses to chemicals to part_of response to X - During the [June 2014 meeting on annotation extension relations](http://wiki.geneontology.org/index.php/Annotation_Extension_meeting_2014-06-16),
it was agreed to change terms for specific responses to chemicals (see
[item 1 from 140624 annotation call minutes](http://wiki.geneontology.org/index.php/Annotation_Conf._Call,_June_24,_2014#Summary_of_proposed_action_items_from_annotation_extension_relation_meeting)).
Examples:
[chemosensory behavior](http://purl.obolibrary.org/obo/GO_0007635)
is_a 'response to chemical' ->
[chemosensory behavior](http://purl.obolibrary.org/obo/GO_0007635)
part_of 'response to chemical'
Motivating example (seems to be fictional term):
‘proteolysis [involved] in cellular response to drug' could have two has_input relationships:
has_input: drug
has_input: proteolysis target.
The has_inputs work in the individual cases but when combined, how do you know which input is which?
A. The drug isn’t an input to the proteolysis. The proteolysis is part of the cellular response to drug. So 'proteolysis [involved] in cellular response to drug' should be part of cellular response to drug, rather than is_a.
Reported by: dosumis
Original Ticket: [geneontology/ontology-requests/10958](https://sourceforge.net/p/geneontology/ontology-requests/10958)
|
process
|
change specific responses to chemicals to part of response to x during the it was agreed to change terms for specific responses to chemicals see examples is a response to chemical part of response to chemical motivating example seems to be fictional term ‘proteolysis in cellular response to drug could have two has input relationships has input drug has input proteolysis target the has inputs work in the individual cases but when combined how do you know which input is which a the drug isn’t an input to the proteolysis the proteolysis is part of the cellular response to drug so proteolysis in cellular response to drug should be part of cellular response to drug rather than is a reported by dosumis original ticket
| 1
|
2,505
| 5,238,993,129
|
IssuesEvent
|
2017-01-31 08:08:21
|
vimperator/vimperator-labs
|
https://api.github.com/repos/vimperator/vimperator-labs
|
closed
|
Vimperator don't work in Firefox Beta. Not e10s.
|
compatibility
|
##### Issue type:
<!-- Pick one and delete the rest -->
- Version compatibility
##### Version:
```
Vimperator 3.15, Nightly 51.0a2, e10s disabled.
```
##### Description:
Mappings and commandline don't work.
In browser console it's showing that `ReferenceError: tabs is not defined`
https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Errors/Not_defined
|
True
|
Vimperator don't work in Firefox Beta. Not e10s. - ##### Issue type:
<!-- Pick one and delete the rest -->
- Version compatibility
##### Version:
```
Vimperator 3.15, Nightly 51.0a2, e10s disabled.
```
##### Description:
Mappings and commandline don't work.
In browser console it's showing that `ReferenceError: tabs is not defined`
https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Errors/Not_defined
|
non_process
|
vimperator don t work in firefox beta not issue type version compatibility version vimperator nightly disabled description mappings and commandline don t work in browser console it s showing that referenceerror tabs is not defined
| 0
|
20,452
| 6,041,080,163
|
IssuesEvent
|
2017-06-10 20:32:30
|
HopefulLlama/UnitTestSCAD
|
https://api.github.com/repos/HopefulLlama/UnitTestSCAD
|
closed
|
Fix "no-return-assign" issue in src/tester/Assertions.js
|
codeclimate technical
|
Return statement should not contain assignment.
https://codeclimate.com/github/HopefulLlama/UnitTestSCAD/src/tester/Assertions.js#issue_593b221e9711610001000034
|
1.0
|
Fix "no-return-assign" issue in src/tester/Assertions.js - Return statement should not contain assignment.
https://codeclimate.com/github/HopefulLlama/UnitTestSCAD/src/tester/Assertions.js#issue_593b221e9711610001000034
|
non_process
|
fix no return assign issue in src tester assertions js return statement should not contain assignment
| 0
|
14,204
| 17,102,793,953
|
IssuesEvent
|
2021-07-09 13:39:06
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
Status of Bazel 5.0.0-pre.20210604.6
|
P1 release team-XProduct type: process
|
- Expected release date: June 11th
Task list:
- [x] Pick release baseline: d571c8976f279e4d485498d0875c85ae73aecc42 with cherrypicks c367a7e027f351e8ffefabf919563abbb8c7d0e1 260015df1524543d56b0dbff00958320cd58d166
- [x] Create release candidate: https://releases.bazel.build/5.0.0/rolling/5.0.0-pre.20210604.6rc1/index.html
- [x] Check Bazel post-submit: https://buildkite.com/bazel/bazel-bazel/builds/16498
- [x] Check downstream projects: https://buildkite.com/bazel/bazel-at-head-plus-downstream/builds/2065
- [x] Push the release: https://releases.bazel.build/5.0.0/rolling/5.0.0-pre.20210604.6/index.html
- [x] Update the [release page](https://github.com/bazelbuild/bazel/releases/)
|
1.0
|
Status of Bazel 5.0.0-pre.20210604.6 - - Expected release date: June 11th
Task list:
- [x] Pick release baseline: d571c8976f279e4d485498d0875c85ae73aecc42 with cherrypicks c367a7e027f351e8ffefabf919563abbb8c7d0e1 260015df1524543d56b0dbff00958320cd58d166
- [x] Create release candidate: https://releases.bazel.build/5.0.0/rolling/5.0.0-pre.20210604.6rc1/index.html
- [x] Check Bazel post-submit: https://buildkite.com/bazel/bazel-bazel/builds/16498
- [x] Check downstream projects: https://buildkite.com/bazel/bazel-at-head-plus-downstream/builds/2065
- [x] Push the release: https://releases.bazel.build/5.0.0/rolling/5.0.0-pre.20210604.6/index.html
- [x] Update the [release page](https://github.com/bazelbuild/bazel/releases/)
|
process
|
status of bazel pre expected release date june task list pick release baseline with cherrypicks create release candidate check bazel post submit check downstream projects push the release update the
| 1
|
406
| 2,848,763,691
|
IssuesEvent
|
2015-05-30 04:02:19
|
dita-ot/dita-ot
|
https://api.github.com/repos/dita-ot/dita-ot
|
closed
|
Error reported during flagging [DOT 2.0]
|
bug P2 preprocess
|
At some point when publishing DITA content to XHTML using DITA OT 2.0 and the latest Saxon 9.6 I get errors like:
[preprocess_flag] Processing C:\Users\radu_coravu\Documents\userguide-private\DITA\temp\webhelp\topics\author-dita-templates.dita
[preprocess_flag] Error on line 563 of flagImpl.xsl:
[preprocess_flag] XTTE0590: An empty sequence is not allowed as the value of parameter $flag-att-val
[preprocess_flag] at xsl:call-template name="gen-prop" (file:/D:/projects/eXml/frameworks/dita/DITA-OT2.x/xsl/preprocess/flagImpl.xsl#541)
[preprocess_flag] at xsl:call-template name="ext-gen-prop" (file:/D:/projects/eXml/frameworks/dita/DITA-OT2.x/xsl/preprocess/flagImpl.xsl#415)
[preprocess_flag] at xsl:call-template name="ext-getrules" (file:/D:/projects/eXml/frameworks/dita/DITA-OT2.x/xsl/preprocess/flagImpl.xsl#375)
[preprocess_flag] at xsl:call-template name="getrules" (file:/D:/projects/eXml/frameworks/dita/DITA-OT2.x/xsl/preprocess/flagImpl.xsl#85)
[preprocess_flag] at xsl:apply-templates (file:/D:/projects/eXml/frameworks/dita/DITA-OT2.x/xsl/preprocess/flagImpl.xsl#80)
[preprocess_flag] processing /topic
[preprocess_flag] Failed to transform document: An empty sequence is not allowed as the value of parameter $flag-att-val
At some point in that stylesheet there is a parameter defined as being of type xs:string.
<xsl:param name="flag-att-val" as="xs:string"/>
But a value for it is passed from this parameter:
<xsl:param name="props" as="xs:string?"/>
Once you add an occurrence indicator after a type, it becomes a sequence (in our case empty sequence or sequence of a single string):
http://www.w3.org/TR/xpath20/#prod-xpath-OccurrenceIndicator
so it seems that the latest Saxon 9.6 reports an error and no longer automatically converts the sequence to a plain string when it is passed as a parameter.
|
1.0
|
Error reported during flagging [DOT 2.0] - At some point when publishing DITA content to XHTML using DITA OT 2.0 and the latest Saxon 9.6 I get errors like:
[preprocess_flag] Processing C:\Users\radu_coravu\Documents\userguide-private\DITA\temp\webhelp\topics\author-dita-templates.dita
[preprocess_flag] Error on line 563 of flagImpl.xsl:
[preprocess_flag] XTTE0590: An empty sequence is not allowed as the value of parameter $flag-att-val
[preprocess_flag] at xsl:call-template name="gen-prop" (file:/D:/projects/eXml/frameworks/dita/DITA-OT2.x/xsl/preprocess/flagImpl.xsl#541)
[preprocess_flag] at xsl:call-template name="ext-gen-prop" (file:/D:/projects/eXml/frameworks/dita/DITA-OT2.x/xsl/preprocess/flagImpl.xsl#415)
[preprocess_flag] at xsl:call-template name="ext-getrules" (file:/D:/projects/eXml/frameworks/dita/DITA-OT2.x/xsl/preprocess/flagImpl.xsl#375)
[preprocess_flag] at xsl:call-template name="getrules" (file:/D:/projects/eXml/frameworks/dita/DITA-OT2.x/xsl/preprocess/flagImpl.xsl#85)
[preprocess_flag] at xsl:apply-templates (file:/D:/projects/eXml/frameworks/dita/DITA-OT2.x/xsl/preprocess/flagImpl.xsl#80)
[preprocess_flag] processing /topic
[preprocess_flag] Failed to transform document: An empty sequence is not allowed as the value of parameter $flag-att-val
At some point in that stylesheet there is a parameter defined as being of type xs:string.
<xsl:param name="flag-att-val" as="xs:string"/>
But a value for it is passed from this parameter:
<xsl:param name="props" as="xs:string?"/>
Once you add an occurrence indicator after a type, it becomes a sequence (in our case empty sequence or sequence of a single string):
http://www.w3.org/TR/xpath20/#prod-xpath-OccurrenceIndicator
so it seems that the latest Saxon 9.6 reports an error and no longer automatically converts the sequence to a plain string when it is passed as a parameter.
|
process
|
error reported during flagging at some point when publishing dita content to xhtml using dita ot and the latest saxon i get errors like processing c users radu coravu documents userguide private dita temp webhelp topics author dita templates dita error on line of flagimpl xsl an empty sequence is not allowed as the value of parameter flag att val at xsl call template name gen prop file d projects exml frameworks dita dita x xsl preprocess flagimpl xsl at xsl call template name ext gen prop file d projects exml frameworks dita dita x xsl preprocess flagimpl xsl at xsl call template name ext getrules file d projects exml frameworks dita dita x xsl preprocess flagimpl xsl at xsl call template name getrules file d projects exml frameworks dita dita x xsl preprocess flagimpl xsl at xsl apply templates file d projects exml frameworks dita dita x xsl preprocess flagimpl xsl processing topic failed to transform document an empty sequence is not allowed as the value of parameter flag att val at some point in that stylesheet there is a parameter defined as being of type xs string but a value for it is passed from this parameter once you add an occurrence indicator after a type it becomes a sequence in our case empty sequence or sequence of a single string so it seems that the latest saxon reports an error and no longer automatically converts the sequence to a plain string when it is passed as a parameter
| 1
|
8,000
| 11,188,530,197
|
IssuesEvent
|
2020-01-02 05:37:52
|
PHPSocialNetwork/phpfastcache
|
https://api.github.com/repos/PHPSocialNetwork/phpfastcache
|
closed
|
Api Method "deleteItemsByTagsAll()" removes unrelated items
|
8.0 :( Not fixable >_< Working & Scheduled [-_-] In Process
|
When You add multiple items that have their own tag and also share a tag it seems that when you delete an item using `deleteItemsByTagsAll(['sharedTag', 'ownTag'])` that all items that share `'sharedTag'` are also removed.
Following the docs it should only remove items that have multiple tags or all you set for this function, not one or more like deleteItemsByTags() does.
This happens for me on Couchbase buckets.
**To Reproduce**
Steps to reproduce the behavior:
1. Save multiple items with a shared tag and their own tag (like ID)
2. Remove one of the items by it's IDtag and the sharedTag using `deleteItemsByTagsAll(['sharedTag', 'ownTag'])`
So 10 items should be in total 21 => 10 items, 10 unique tags items, 1 shared tag item
**Expected behavior**
One item removed and one (unique) tag, so 2 items less in the bucket and the rest should stay
**Current behavior**
All items that share the shared tags are removed, so also are their individual tags.
|
1.0
|
Api Method "deleteItemsByTagsAll()" removes unrelated items - When You add multiple items that have their own tag and also share a tag it seems that when you delete an item using `deleteItemsByTagsAll(['sharedTag', 'ownTag'])` that all items that share `'sharedTag'` are also removed.
Following the docs it should only remove items that have multiple tags or all you set for this function, not one or more like deleteItemsByTags() does.
This happens for me on Couchbase buckets.
**To Reproduce**
Steps to reproduce the behavior:
1. Save multiple items with a shared tag and their own tag (like ID)
2. Remove one of the items by it's IDtag and the sharedTag using `deleteItemsByTagsAll(['sharedTag', 'ownTag'])`
So 10 items should be in total 21 => 10 items, 10 unique tags items, 1 shared tag item
**Expected behavior**
One item removed and one (unique) tag, so 2 items less in the bucket and the rest should stay
**Current behavior**
All items that share the shared tags are removed, so also are their individual tags.
|
process
|
api method deleteitemsbytagsall removes unrelated items when you add multiple items that have their own tag and also share a tag it seems that when you delete an item using deleteitemsbytagsall that all items that share sharedtag are also removed following the docs it should only remove items that have multiple tags or all you set for this function not one or more like deleteitemsbytags does this happens for me on couchbase buckets to reproduce steps to reproduce the behavior save multiple items with a shared tag and their own tag like id remove one of the items by it s idtag and the sharedtag using deleteitemsbytagsall so items should be in total items unique tags items shared tag item expected behavior one item removed and one unique tag so items less in the bucket and the rest should stay current behavior all items that share the shared tags are removed so also are their individual tags
| 1
|
20,074
| 26,568,442,866
|
IssuesEvent
|
2023-01-20 23:09:08
|
srophe/caesarea-data
|
https://api.github.com/repos/srophe/caesarea-data
|
closed
|
Test out why pipe character conversion to <lb/> isn't working consistently
|
post-processor
|
Cf. #150. It seems like it only worked part of the time. Need to dig into this a bit more.
|
1.0
|
Test out why pipe character conversion to <lb/> isn't working consistently - Cf. #150. It seems like it only worked part of the time. Need to dig into this a bit more.
|
process
|
test out why pipe character conversion to isn t working consistently cf it seems like it only worked part of the time need to dig into this a bit more
| 1
|
11,632
| 14,490,871,169
|
IssuesEvent
|
2020-12-11 03:17:05
|
cypress-io/cypress
|
https://api.github.com/repos/cypress-io/cypress
|
reopened
|
"Found .only in folder" error when trying to commit
|
process: contributing type: chore
|
### Current behavior:
I tried to commit into my own branch in my fork of cypress. This fails with the error:
> Found .only in folder(s) 👎
> 544: * Mocks global clock but only overrides specific functions.
> 1847: * Creates a spy / stub but only for calls with given arguments.
> 2108: * Whether Cypress will upload the video to the Dashboard even if all tests are passing. This applies only when recording your runs to the Dashboard. Turn this off if you’d like the video uploaded only when there are failing tests.
> 2721: * Causes all `.keys` assertions that follow in the chain to only require that the target have at least one of the given keys. This is the opposite of `.all`, which requires that the target have all of the given keys.
> 3221: * Causes all `.keys` assertions that follow in the chain to only require that the target not have at least one of the given keys. This is the opposite of `.all`, which requires that the target have all of the given keys.
> 4352: /** Override *name* for display purposes only */
But I have done nothing related to this I think.
### Desired behavior:
I should be able to commit on Windows without using "--no-verify"
### Steps to reproduce: (app code and test code)
Fork cypress-io/cypress repo, make a branch and try to commit the changes made in https://github.com/gabbersepp/cypress/commit/de0a7c0d4936886b1b7576b60bb0b533c9207cb2
### Versions
Current develop branch, Win10
|
1.0
|
"Found .only in folder" error when trying to commit - ### Current behavior:
I tried to commit into my own branch in my fork of cypress. This fails with the error:
> Found .only in folder(s) 👎
> 544: * Mocks global clock but only overrides specific functions.
> 1847: * Creates a spy / stub but only for calls with given arguments.
> 2108: * Whether Cypress will upload the video to the Dashboard even if all tests are passing. This applies only when recording your runs to the Dashboard. Turn this off if you’d like the video uploaded only when there are failing tests.
> 2721: * Causes all `.keys` assertions that follow in the chain to only require that the target have at least one of the given keys. This is the opposite of `.all`, which requires that the target have all of the given keys.
> 3221: * Causes all `.keys` assertions that follow in the chain to only require that the target not have at least one of the given keys. This is the opposite of `.all`, which requires that the target have all of the given keys.
> 4352: /** Override *name* for display purposes only */
But I have done nothing related to this I think.
### Desired behavior:
I should be able to commit on Windows without using "--no-verify"
### Steps to reproduce: (app code and test code)
Fork cypress-io/cypress repo, make a branch and try to commit the changes made in https://github.com/gabbersepp/cypress/commit/de0a7c0d4936886b1b7576b60bb0b533c9207cb2
### Versions
Current develop branch, Win10
|
process
|
found only in folder error when trying to commit current behavior i tried to commit into my own branch in my fork of cypress this fails with the error found only in folder s 👎 mocks global clock but only overrides specific functions creates a spy stub but only for calls with given arguments whether cypress will upload the video to the dashboard even if all tests are passing this applies only when recording your runs to the dashboard turn this off if you’d like the video uploaded only when there are failing tests causes all keys assertions that follow in the chain to only require that the target have at least one of the given keys this is the opposite of all which requires that the target have all of the given keys causes all keys assertions that follow in the chain to only require that the target not have at least one of the given keys this is the opposite of all which requires that the target have all of the given keys override name for display purposes only but i have done nothing related to this i think desired behavior i should be able to commit on windows without using no verify steps to reproduce app code and test code fork cypress io cypress repo make a branch and try to commit the changes made in versions current develop branch
| 1
|
9,499
| 12,488,728,271
|
IssuesEvent
|
2020-05-31 15:33:03
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
closed
|
Sorting by two columns with the same name fails with ambiguous column error
|
Priority:P2 Querying/Processor Type:Bug
|
For example if you create a query that has two `count` columns in the results and then click the column header to sort the resulting query will fail because the sort is ambiguous. This can be confusing because it's not obvious how to remove sorts in the query builder (you need to use the `...` menu even though you didn't need to use it to add the sort).
There's a couple things we could do to fix this:
1. disallow sorts for ambiguous columns since we know they'll break
2. try to prevent ambiguous columns from being created in the first place. e.g. by naming our columns `count` and `count_2` in the query instead of having `clojure.java.jdbc` do it for us in the results.
3. let people manually specify column names, perhaps just reusing the existing UI for choosing which columns to display/etc.
|
1.0
|
Sorting by two columns with the same name fails with ambiguous column error - For example if you create a query that has two `count` columns in the results and then click the column header to sort the resulting query will fail because the sort is ambiguous. This can be confusing because it's not obvious how to remove sorts in the query builder (you need to use the `...` menu even though you didn't need to use it to add the sort).
There's a couple things we could do to fix this:
1. disallow sorts for ambiguous columns since we know they'll break
2. try to prevent ambiguous columns from being created in the first place. e.g. by naming our columns `count` and `count_2` in the query instead of having `clojure.java.jdbc` do it for us in the results.
3. let people manually specify column names, perhaps just reusing the existing UI for choosing which columns to display/etc.
|
process
|
sorting by two columns with the same name fails with ambiguous column error for example if you create a query that has two count columns in the results and then click the column header to sort the resulting query will fail because the sort is ambiguous this can be confusing because it s not obvious how to remove sorts in the query builder you need to use the menu even though you didn t need to use it to add the sort there s a couple things we could do to fix this disallow sorts for ambiguous columns since we know they ll break try to prevent ambiguous columns from being created in the first place e g by naming our columns count and count in the query instead of having clojure java jdbc do it for us in the results let people manually specify column names perhaps just reusing the existing ui for choosing which columns to display etc
| 1
|
127,515
| 27,063,260,597
|
IssuesEvent
|
2023-02-13 21:36:53
|
eclipse-theia/theia
|
https://api.github.com/repos/eclipse-theia/theia
|
closed
|
Infer plugin activation events from contributions
|
plug-in system vscode
|
### Feature Description:
Starting with the newest version of VSCode, the plugin host infers activation events in part from the `package.json` contributions. This is highlighted by a warning in the activation events:

We should implement a similar mechanism soon, as otherwise newer extensions (with old vscode version requirements) might not be activated as expected.
|
1.0
|
Infer plugin activation events from contributions - ### Feature Description:
Starting with the newest version of VSCode, the plugin host infers activation events in part from the `package.json` contributions. This is highlighted by a warning in the activation events:

We should implement a similar mechanism soon, as otherwise newer extensions (with old vscode version requirements) might not be activated as expected.
|
non_process
|
infer plugin activation events from contributions feature description starting with the newest version of vscode the plugin host infers activation events in part from the package json contributions this is highlighted by a warning in the activation events we should implement a similar mechanism soon as otherwise newer extensions with old vscode version requirements might not be activated as expected
| 0
|
971
| 2,522,283,567
|
IssuesEvent
|
2015-01-19 20:50:42
|
vanstyn/RapidApp
|
https://api.github.com/repos/vanstyn/RapidApp
|
opened
|
Unit tests for SimpleCAS/CMS uploads
|
Test Coverage
|
Using ``RapidApp::Test``, unit tests need to be written to confirm operation of file/image uploads:
* Uploading a normal file and getting back link
* Access the uploaded file via that link
* Uploading an image and getting back link+meta data info (height/width)
* Uploading an image that is larger than max size and getting back warning and resized details
|
1.0
|
Unit tests for SimpleCAS/CMS uploads - Using ``RapidApp::Test``, unit tests need to be written to confirm operation of file/image uploads:
* Uploading a normal file and getting back link
* Access the uploaded file via that link
* Uploading an image and getting back link+meta data info (height/width)
* Uploading an image that is larger than max size and getting back warning and resized details
|
non_process
|
unit tests for simplecas cms uploads using rapidapp test unit tests need to be written to confirm operation of file image uploads uploading a normal file and getting back link access the uploaded file via that link uploading an image and getting back link meta data info height width uploading an image that is larger than max size and getting back warning and resized details
| 0
|
105,323
| 11,441,158,043
|
IssuesEvent
|
2020-02-05 11:04:54
|
IlonaMenkui/super-enigma-service
|
https://api.github.com/repos/IlonaMenkui/super-enigma-service
|
closed
|
Write Swagger spec
|
documentation enhancement
|
Write Swagger spec about available REST endpoints. Place this doc under `/spec/http/api-v1.yml` folder.
Resources:
1. Doc - https://swagger.io/docs/specification/2-0/basic-structure/
2. Examples - https://github.com/OAI/OpenAPI-Specification/tree/master/examples/v2.0/yaml
|
1.0
|
Write Swagger spec - Write Swagger spec about available REST endpoints. Place this doc under `/spec/http/api-v1.yml` folder.
Resources:
1. Doc - https://swagger.io/docs/specification/2-0/basic-structure/
2. Examples - https://github.com/OAI/OpenAPI-Specification/tree/master/examples/v2.0/yaml
|
non_process
|
write swagger spec write swagger spec about available rest endpoints place this doc under spec http api yml folder resources doc examples
| 0
|
19,626
| 25,981,257,693
|
IssuesEvent
|
2022-12-19 19:04:13
|
googleapis/google-cloud-go
|
https://api.github.com/repos/googleapis/google-cloud-go
|
opened
|
logging: use custom environment test container
|
api: logging type: process priority: p3
|
The logging environment tests are currently running using the same container base as the other go tests. Sometimes, something in the base image shifts, and causes test failures. We should consider creating a new container base image specifically for environment tests.
Doing this, we can bake more dependencies into the base image: https://github.com/googleapis/google-cloud-go/pull/7159#discussion_r1052536825
|
1.0
|
logging: use custom environment test container - The logging environment tests are currently running using the same container base as the other go tests. Sometimes, something in the base image shifts, and causes test failures. We should consider creating a new container base image specifically for environment tests.
Doing this, we can bake more dependencies into the base image: https://github.com/googleapis/google-cloud-go/pull/7159#discussion_r1052536825
|
process
|
logging use custom environment test container the logging environment tests are currently running using the same container base as the other go tests sometimes something in the base image shifts and causes test failures we should consider creating a new container base image specifically for environment tests doing this we can bake more dependencies into the base image
| 1
|
19,748
| 26,109,477,253
|
IssuesEvent
|
2022-12-27 17:30:57
|
SigNoz/signoz-otel-collector
|
https://api.github.com/repos/SigNoz/signoz-otel-collector
|
closed
|
Check impact of keeping cache size to 1M or 100M in signozspanmetricsprocessor
|
signozspanmetricsprocessor
|
<a href="https://github.com/ankitnayan"><img src="https://avatars.githubusercontent.com/u/12460410?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [ankitnayan](https://github.com/ankitnayan)**
_Monday Jul 11, 2022 at 12:09 GMT_
_Originally opened as https://github.com/SigNoz/opentelemetry-collector-contrib/issues/1851_
----
It is not possible to determine the cache size beforehand as users keep on adding services to signoz. We should have an idea of the impact of keeping a high number for `dimensions_cache_size` like `1M` or `100M`. Or should we change the size dynamically based on some metric/number?
|
1.0
|
Check impact of keeping cache size to 1M or 100M in signozspanmetricsprocessor - <a href="https://github.com/ankitnayan"><img src="https://avatars.githubusercontent.com/u/12460410?v=4" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [ankitnayan](https://github.com/ankitnayan)**
_Monday Jul 11, 2022 at 12:09 GMT_
_Originally opened as https://github.com/SigNoz/opentelemetry-collector-contrib/issues/1851_
----
It is not possible to determine the cache size beforehand as users keep on adding services to signoz. We should have an idea of the impact of keeping a high number for `dimensions_cache_size` like `1M` or `100M`. Or should we change the size dynamically based on some metric/number?
|
process
|
check impact of keeping cache size to or in signozspanmetricsprocessor issue by monday jul at gmt originally opened as it is not possible to determine the cache size beforehand as users keep on adding services to signoz we should have an idea of the impact of keeping a high number for dimensions cache size like or or should we change the size dynamically based on some metric number
| 1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.