Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 7
112
| repo_url
stringlengths 36
141
| action
stringclasses 3
values | title
stringlengths 1
744
| labels
stringlengths 4
574
| body
stringlengths 9
211k
| index
stringclasses 10
values | text_combine
stringlengths 96
211k
| label
stringclasses 2
values | text
stringlengths 96
188k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
19,392
| 25,534,378,949
|
IssuesEvent
|
2022-11-29 10:49:35
|
inmanta/web-console
|
https://api.github.com/repos/inmanta/web-console
|
opened
|
Improved cleanup on Jenkins pipline
|
process
|
The current Jenkinsfile executes the following commands in the `Testing with cypress` stage of the pipline:
```
yarn run setup-server:lsm;
yarn run cypress-test;
yarn run kill-server:lsm
```
The script will stop executing as soon as one command fails. This means that when a test case fails, `yarn run kill-server:lsm` will not get executed because `yarn run cypress-test` has exited with a non-zero exit code. It would be good to move the `yarn run kill-server:lsm` command to the `cleanup` post build step of the pipeline to ensure that a cleanup happens in all circumstances.
|
1.0
|
Improved cleanup on Jenkins pipline - The current Jenkinsfile executes the following commands in the `Testing with cypress` stage of the pipline:
```
yarn run setup-server:lsm;
yarn run cypress-test;
yarn run kill-server:lsm
```
The script will stop executing as soon as one command fails. This means that when a test case fails, `yarn run kill-server:lsm` will not get executed because `yarn run cypress-test` has exited with a non-zero exit code. It would be good to move the `yarn run kill-server:lsm` command to the `cleanup` post build step of the pipeline to ensure that a cleanup happens in all circumstances.
|
process
|
improved cleanup on jenkins pipline the current jenkinsfile executes the following commands in the testing with cypress stage of the pipline yarn run setup server lsm yarn run cypress test yarn run kill server lsm the script will stop executing as soon as one command fails this means that when a test case fails yarn run kill server lsm will not get executed because yarn run cypress test has exited with a non zero exit code it would be good to move the yarn run kill server lsm command to the cleanup post build step of the pipeline to ensure that a cleanup happens in all circumstances
| 1
|
12,014
| 14,738,399,412
|
IssuesEvent
|
2021-01-07 04:39:07
|
kdjstudios/SABillingGitlab
|
https://api.github.com/repos/kdjstudios/SABillingGitlab
|
closed
|
Structured Releases
|
anc-process anp-important
|
In GitLab by @kdjstudios on May 18, 2018, 09:03
@smasih @pchaudhary @tim.traylor @thomas.gentry
Hello Team,
I am needing to know which server we are using for the primary releases to Production?
I will be making all Help Tool Tips, News Feed, Wiki Help, and other updates to this one server then that will get released along with the other updates on our bi weekly schedule.
|
1.0
|
Structured Releases - In GitLab by @kdjstudios on May 18, 2018, 09:03
@smasih @pchaudhary @tim.traylor @thomas.gentry
Hello Team,
I am needing to know which server we are using for the primary releases to Production?
I will be making all Help Tool Tips, News Feed, Wiki Help, and other updates to this one server then that will get released along with the other updates on our bi weekly schedule.
|
process
|
structured releases in gitlab by kdjstudios on may smasih pchaudhary tim traylor thomas gentry hello team i am needing to know which server we are using for the primary releases to production i will be making all help tool tips news feed wiki help and other updates to this one server then that will get released along with the other updates on our bi weekly schedule
| 1
|
18,783
| 24,690,287,548
|
IssuesEvent
|
2022-10-19 08:02:42
|
geneontology/go-ontology
|
https://api.github.com/repos/geneontology/go-ontology
|
closed
|
homeostatic process (GO:0042592) parentage query
|
cellular processes
|
I have been wondering why
homeostatic process (GO:0042592)
Any biological process involved in the maintenance of an internal steady state.
is a descendant of
regulation of biological quality (GO:0065008)
Any process that modulates a qualitative or quantitative trait of a biological quality. A biological quality is a measurable attribute of an organism or part of an organism, such as size, mass, shape, color, etc.
homeostasis terms are problematic anyway, because a lot of regulation *is*. homeostasis, but also not all of homeostasis is regulation.
So we probably need to really think about what we are trying to capture with homeostasis. For example it might have more value for metabolic processes, but it would be very difficult to be precise about when a pathway is part of homeostasis. Int the example of. "zinc transport" some transporters transport since whenever they detect it, but some pathways are activated to maintain the correct concentrations in organelles (as far as I understand it).
In the short term, so that we can properly assess which genes are annotated to "regulation of process" terms, it might be useful to remove the "regulation" parent from cellular homeostasis?
|
1.0
|
homeostatic process (GO:0042592) parentage query -
I have been wondering why
homeostatic process (GO:0042592)
Any biological process involved in the maintenance of an internal steady state.
is a descendant of
regulation of biological quality (GO:0065008)
Any process that modulates a qualitative or quantitative trait of a biological quality. A biological quality is a measurable attribute of an organism or part of an organism, such as size, mass, shape, color, etc.
homeostasis terms are problematic anyway, because a lot of regulation *is*. homeostasis, but also not all of homeostasis is regulation.
So we probably need to really think about what we are trying to capture with homeostasis. For example it might have more value for metabolic processes, but it would be very difficult to be precise about when a pathway is part of homeostasis. Int the example of. "zinc transport" some transporters transport since whenever they detect it, but some pathways are activated to maintain the correct concentrations in organelles (as far as I understand it).
In the short term, so that we can properly assess which genes are annotated to "regulation of process" terms, it might be useful to remove the "regulation" parent from cellular homeostasis?
|
process
|
homeostatic process go parentage query i have been wondering why homeostatic process go any biological process involved in the maintenance of an internal steady state is a descendant of regulation of biological quality go any process that modulates a qualitative or quantitative trait of a biological quality a biological quality is a measurable attribute of an organism or part of an organism such as size mass shape color etc homeostasis terms are problematic anyway because a lot of regulation is homeostasis but also not all of homeostasis is regulation so we probably need to really think about what we are trying to capture with homeostasis for example it might have more value for metabolic processes but it would be very difficult to be precise about when a pathway is part of homeostasis int the example of zinc transport some transporters transport since whenever they detect it but some pathways are activated to maintain the correct concentrations in organelles as far as i understand it in the short term so that we can properly assess which genes are annotated to regulation of process terms it might be useful to remove the regulation parent from cellular homeostasis
| 1
|
5,648
| 8,513,547,837
|
IssuesEvent
|
2018-10-31 16:18:07
|
FaruqRahmadani/Presensi-Guru-2
|
https://api.github.com/repos/FaruqRahmadani/Presensi-Guru-2
|
closed
|
Variabel sass useless
|
PROCESS
|
### File : resources/assets/sass/_mentah/_variable.scss
> 
### File :resources/assets/sass/_mentah/_login.scss
> 
### Permasalahan
asaku sayang mun kita sdh declare variabel tp kd dipakai.
|
1.0
|
Variabel sass useless - ### File : resources/assets/sass/_mentah/_variable.scss
> 
### File :resources/assets/sass/_mentah/_login.scss
> 
### Permasalahan
asaku sayang mun kita sdh declare variabel tp kd dipakai.
|
process
|
variabel sass useless file resources assets sass mentah variable scss file resources assets sass mentah login scss permasalahan asaku sayang mun kita sdh declare variabel tp kd dipakai
| 1
|
6,099
| 8,958,406,426
|
IssuesEvent
|
2019-01-27 14:05:01
|
lxqt/lxqt
|
https://api.github.com/repos/lxqt/lxqt
|
closed
|
Notification icon doesn't update
|
wont-process-this wontfix
|
When an application shows notifications the first image is used for all said app's notifications.
This is a problem when something like spotify is showing what song is being played, as it will only show the album art of the first song played for all future songs - presumably it's cached on first display.
Version: 0.13
|
1.0
|
Notification icon doesn't update - When an application shows notifications the first image is used for all said app's notifications.
This is a problem when something like spotify is showing what song is being played, as it will only show the album art of the first song played for all future songs - presumably it's cached on first display.
Version: 0.13
|
process
|
notification icon doesn t update when an application shows notifications the first image is used for all said app s notifications this is a problem when something like spotify is showing what song is being played as it will only show the album art of the first song played for all future songs presumably it s cached on first display version
| 1
|
17,961
| 23,970,061,130
|
IssuesEvent
|
2022-09-13 06:55:56
|
cypress-io/cypress
|
https://api.github.com/repos/cypress-io/cypress
|
opened
|
[ACI] Investigate better testing strategy
|
process: tests stage: internal
|
### What would you like?
Testing ACI PRs is [hard and time consuming](https://github.com/cypress-io/cypress/pull/23735#issuecomment-1243296033), since a lot of the logic is "after X days, Y should happen" which is based on data in the global cache (read from user's machine). See this PR https://github.com/cypress-io/cypress/pull/23735 for an example of the complexity to test manually.
I think we can automate most of this. The gaps in our current infra (within the monorepo) are:
1. no examples on how to update the cache (the file on disc persisted between app sessions)
2. (probably) no endpoint to refetch this (I think we just fetch it once, on startup)
If we could implement these things, we could probably automate about 90% of the test plan in the PR linked above. We should do this.
### Why is this needed?
_No response_
### Other
_No response_
|
1.0
|
[ACI] Investigate better testing strategy - ### What would you like?
Testing ACI PRs is [hard and time consuming](https://github.com/cypress-io/cypress/pull/23735#issuecomment-1243296033), since a lot of the logic is "after X days, Y should happen" which is based on data in the global cache (read from user's machine). See this PR https://github.com/cypress-io/cypress/pull/23735 for an example of the complexity to test manually.
I think we can automate most of this. The gaps in our current infra (within the monorepo) are:
1. no examples on how to update the cache (the file on disc persisted between app sessions)
2. (probably) no endpoint to refetch this (I think we just fetch it once, on startup)
If we could implement these things, we could probably automate about 90% of the test plan in the PR linked above. We should do this.
### Why is this needed?
_No response_
### Other
_No response_
|
process
|
investigate better testing strategy what would you like testing aci prs is since a lot of the logic is after x days y should happen which is based on data in the global cache read from user s machine see this pr for an example of the complexity to test manually i think we can automate most of this the gaps in our current infra within the monorepo are no examples on how to update the cache the file on disc persisted between app sessions probably no endpoint to refetch this i think we just fetch it once on startup if we could implement these things we could probably automate about of the test plan in the pr linked above we should do this why is this needed no response other no response
| 1
|
11,157
| 13,957,693,713
|
IssuesEvent
|
2020-10-24 08:11:03
|
alexanderkotsev/geoportal
|
https://api.github.com/repos/alexanderkotsev/geoportal
|
opened
|
BE: Missing resources in Geoportal
|
BE - Belgium Geoportal Harvesting process
|
Collected from the Geoportal Workshop online survey answers:
At Walloon Region level:
- Geoportal detects the link to VS and DS correclty.
- Regarding the unique resource identifier, Geoportal doesn’t detect the code and Namespace declared in the metadata cause to the fact we declare two Unique ID:
i) the global ID using gmd:citation/gmd:CI_Citation/gmd:identifier/*/gmd:code element and its sub-type “RS_Identifier” which is linked to the dataset and
ii) the local ID using gmd:citation/gmd:CI_Citation/gmd:identifier/*/gmd:code element and its sub-type “MD_Identifier” which is a characterString referring the short name of the dataset in our internal database.
http://inspire-geoportal.ec.europa.eu/download_details.html?view=downloadDetails&resourceId=%2FINSPIRE-f5201eb0-2404-11e5-8130-52540004b857_20181206-123520%2Fservices%2F1%2FPullResults%2F141-160%2Fseries%2F7&expandedSection=metadata
|
1.0
|
BE: Missing resources in Geoportal - Collected from the Geoportal Workshop online survey answers:
At Walloon Region level:
- Geoportal detects the link to VS and DS correclty.
- Regarding the unique resource identifier, Geoportal doesn’t detect the code and Namespace declared in the metadata cause to the fact we declare two Unique ID:
i) the global ID using gmd:citation/gmd:CI_Citation/gmd:identifier/*/gmd:code element and its sub-type “RS_Identifier” which is linked to the dataset and
ii) the local ID using gmd:citation/gmd:CI_Citation/gmd:identifier/*/gmd:code element and its sub-type “MD_Identifier” which is a characterString referring the short name of the dataset in our internal database.
http://inspire-geoportal.ec.europa.eu/download_details.html?view=downloadDetails&resourceId=%2FINSPIRE-f5201eb0-2404-11e5-8130-52540004b857_20181206-123520%2Fservices%2F1%2FPullResults%2F141-160%2Fseries%2F7&expandedSection=metadata
|
process
|
be missing resources in geoportal collected from the geoportal workshop online survey answers at walloon region level geoportal detects the link to vs and ds correclty regarding the unique resource identifier geoportal doesn rsquo t detect the code and namespace declared in the metadata cause to the fact we declare two unique id i the global id using gmd citation gmd ci citation gmd identifier gmd code element and its sub type ldquo rs identifier rdquo which is linked to the dataset and ii the local id using gmd citation gmd ci citation gmd identifier gmd code element and its sub type ldquo md identifier rdquo which is a characterstring referring the short name of the dataset in our internal database
| 1
|
12,360
| 14,888,288,181
|
IssuesEvent
|
2021-01-20 19:37:09
|
encode/uvicorn
|
https://api.github.com/repos/encode/uvicorn
|
closed
|
Huge number of workers fired up when I tried to run the app using PyInstaller generated byte code
|
multiprocessing need confirmation
|
### Checklist
<!-- Please make sure you check all these items before submitting your bug report. -->
- [Yes ] The bug is reproducible against the latest release and/or `master`. [Tested using uvicorn version 0.12.1]
- [Yes] There are no similar issues or pull requests to fix it yet.
### Describe the bug
<!-- A clear and concise description of what the bug is. -->
We bundled our app using pyinstaller byte code. When we try to use 4 workers to run the application, a huge number of workers are fired up resulting in our server going down. When we remove the "workers = 4" parameter in the Uvicorn.run(), everything is working fine.
### To reproduce
<!-- Provide a *minimal* example with steps to reproduce the bug locally.
NOTE: try to keep any external dependencies *at an absolute minimum* .
In other words, remove anything that doesn't make the bug go away.
-->
Server: RHEL Server on AWS EC2 instance
When I generated the pyinstaller byte code using the below files, everything works fine.
"test.py" and "test.spec" files used to generate the byte code executable.
Steps:
1. Install PyInstaller and run "pyinstaller --onefile -w test.spec". A byte code executable 'test' is generated in dist/ directory.
2. Run the executable as './test'.
[testspec.txt](https://github.com/encode/uvicorn/files/5839125/testspec.txt)
[testpy.txt](https://github.com/encode/uvicorn/files/5839126/testpy.txt)
<img width="560" alt="Screen Shot 2021-01-19 at 5 58 39 PM" src="https://user-images.githubusercontent.com/32053777/105105227-a9253100-5a81-11eb-884c-fa6fb1f187de.png">
But when we set the number of workers to 4, huge number of workers are fired up resulting in server going down.
[test1spec.txt](https://github.com/encode/uvicorn/files/5839127/test1spec.txt)
[test1py.txt](https://github.com/encode/uvicorn/files/5839128/test1py.txt)
Steps:
Repeat the same procedure as above with test1.spec.
<img width="537" alt="Screen Shot 2021-01-19 at 5 45 52 PM" src="https://user-images.githubusercontent.com/32053777/105105312-deca1a00-5a81-11eb-91db-e9b344902393.png">
### Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
Expected 4 workers to start.
### Actual behavior
<!-- A clear and concise description of what actually happens. -->
Huge number of workers are fired up resulting in server going down.
### Debugging material
<!-- Any tracebacks, screenshots, etc. that can help understanding the problem.
NOTE:
- Please list tracebacks in full (don't truncate them).
- If relevant, consider turning on DEBUG or TRACE logs for additional details (see the Logging section on https://www.uvicorn.org/settings/ specifically the `log-level` flag).
- Consider using `<details>` to make tracebacks/logs collapsible if they're very large (see https://gist.github.com/ericclemmons/b146fe5da72ca1f706b2ef72a20ac39d).
-->
### Environment
OS: RHEL Server on AWS EC2 instance
Python: Python 3.6.8
Uvicorn version: uvicorn 0.12.1 with CPython 3.8.3 on Linux
- uvicorn.run('test:app', host='0.0.0.0', workers=4, port=7500)
### Additional context
<!-- Any additional information that can help understanding the problem.
Eg. linked issues, or a description of what you were trying to achieve. -->
|
1.0
|
Huge number of workers fired up when I tried to run the app using PyInstaller generated byte code - ### Checklist
<!-- Please make sure you check all these items before submitting your bug report. -->
- [Yes ] The bug is reproducible against the latest release and/or `master`. [Tested using uvicorn version 0.12.1]
- [Yes] There are no similar issues or pull requests to fix it yet.
### Describe the bug
<!-- A clear and concise description of what the bug is. -->
We bundled our app using pyinstaller byte code. When we try to use 4 workers to run the application, a huge number of workers are fired up resulting in our server going down. When we remove the "workers = 4" parameter in the Uvicorn.run(), everything is working fine.
### To reproduce
<!-- Provide a *minimal* example with steps to reproduce the bug locally.
NOTE: try to keep any external dependencies *at an absolute minimum* .
In other words, remove anything that doesn't make the bug go away.
-->
Server: RHEL Server on AWS EC2 instance
When I generated the pyinstaller byte code using the below files, everything works fine.
"test.py" and "test.spec" files used to generate the byte code executable.
Steps:
1. Install PyInstaller and run "pyinstaller --onefile -w test.spec". A byte code executable 'test' is generated in dist/ directory.
2. Run the executable as './test'.
[testspec.txt](https://github.com/encode/uvicorn/files/5839125/testspec.txt)
[testpy.txt](https://github.com/encode/uvicorn/files/5839126/testpy.txt)
<img width="560" alt="Screen Shot 2021-01-19 at 5 58 39 PM" src="https://user-images.githubusercontent.com/32053777/105105227-a9253100-5a81-11eb-884c-fa6fb1f187de.png">
But when we set the number of workers to 4, huge number of workers are fired up resulting in server going down.
[test1spec.txt](https://github.com/encode/uvicorn/files/5839127/test1spec.txt)
[test1py.txt](https://github.com/encode/uvicorn/files/5839128/test1py.txt)
Steps:
Repeat the same procedure as above with test1.spec.
<img width="537" alt="Screen Shot 2021-01-19 at 5 45 52 PM" src="https://user-images.githubusercontent.com/32053777/105105312-deca1a00-5a81-11eb-91db-e9b344902393.png">
### Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
Expected 4 workers to start.
### Actual behavior
<!-- A clear and concise description of what actually happens. -->
Huge number of workers are fired up resulting in server going down.
### Debugging material
<!-- Any tracebacks, screenshots, etc. that can help understanding the problem.
NOTE:
- Please list tracebacks in full (don't truncate them).
- If relevant, consider turning on DEBUG or TRACE logs for additional details (see the Logging section on https://www.uvicorn.org/settings/ specifically the `log-level` flag).
- Consider using `<details>` to make tracebacks/logs collapsible if they're very large (see https://gist.github.com/ericclemmons/b146fe5da72ca1f706b2ef72a20ac39d).
-->
### Environment
OS: RHEL Server on AWS EC2 instance
Python: Python 3.6.8
Uvicorn version: uvicorn 0.12.1 with CPython 3.8.3 on Linux
- uvicorn.run('test:app', host='0.0.0.0', workers=4, port=7500)
### Additional context
<!-- Any additional information that can help understanding the problem.
Eg. linked issues, or a description of what you were trying to achieve. -->
|
process
|
huge number of workers fired up when i tried to run the app using pyinstaller generated byte code checklist the bug is reproducible against the latest release and or master there are no similar issues or pull requests to fix it yet describe the bug we bundled our app using pyinstaller byte code when we try to use workers to run the application a huge number of workers are fired up resulting in our server going down when we remove the workers parameter in the uvicorn run everything is working fine to reproduce provide a minimal example with steps to reproduce the bug locally note try to keep any external dependencies at an absolute minimum in other words remove anything that doesn t make the bug go away server rhel server on aws instance when i generated the pyinstaller byte code using the below files everything works fine test py and test spec files used to generate the byte code executable steps install pyinstaller and run pyinstaller onefile w test spec a byte code executable test is generated in dist directory run the executable as test img width alt screen shot at pm src but when we set the number of workers to huge number of workers are fired up resulting in server going down steps repeat the same procedure as above with spec img width alt screen shot at pm src expected behavior expected workers to start actual behavior huge number of workers are fired up resulting in server going down debugging material any tracebacks screenshots etc that can help understanding the problem note please list tracebacks in full don t truncate them if relevant consider turning on debug or trace logs for additional details see the logging section on specifically the log level flag consider using to make tracebacks logs collapsible if they re very large see environment os rhel server on aws instance python python uvicorn version uvicorn with cpython on linux uvicorn run test app host workers port additional context any additional information that can help understanding the problem eg linked issues or a description of what you were trying to achieve
| 1
|
86,953
| 17,108,289,047
|
IssuesEvent
|
2021-07-09 22:04:39
|
phetsims/circuit-construction-kit-common
|
https://api.github.com/repos/phetsims/circuit-construction-kit-common
|
closed
|
What is the convention for naming imported sounds?
|
dev:code-review
|
In code review https://github.com/phetsims/circuit-construction-kit-common/issues/672 @jonathanolson asked:
```js
//REVIEW: What is the convention for naming of these?
import dogBarkSound from '../../sounds/dog-bark_mp3.js';
```
|
1.0
|
What is the convention for naming imported sounds? - In code review https://github.com/phetsims/circuit-construction-kit-common/issues/672 @jonathanolson asked:
```js
//REVIEW: What is the convention for naming of these?
import dogBarkSound from '../../sounds/dog-bark_mp3.js';
```
|
non_process
|
what is the convention for naming imported sounds in code review jonathanolson asked js review what is the convention for naming of these import dogbarksound from sounds dog bark js
| 0
|
20,413
| 27,072,268,080
|
IssuesEvent
|
2023-02-14 08:01:47
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
Open source internal version of BlazeJavacMainTest
|
P4 type: process team-Rules-Java stale
|
After https://bazel-review.git.corp.google.com/c/10751/ is submitted there will be 2 versions of this test (internal + external), which is not desirable. Open sourcing the internal version also improves contributions test coverage.
|
1.0
|
Open source internal version of BlazeJavacMainTest - After https://bazel-review.git.corp.google.com/c/10751/ is submitted there will be 2 versions of this test (internal + external), which is not desirable. Open sourcing the internal version also improves contributions test coverage.
|
process
|
open source internal version of blazejavacmaintest after is submitted there will be versions of this test internal external which is not desirable open sourcing the internal version also improves contributions test coverage
| 1
|
176,476
| 14,583,616,295
|
IssuesEvent
|
2020-12-18 14:16:05
|
ChainYo/StableMatching
|
https://api.github.com/repos/ChainYo/StableMatching
|
closed
|
Update Readme
|
documentation good first issue
|
We need to update [Readme.md] to keep informations up to date.
- [x] Quick presentation for the project
- [x] Explanation of the goal of the project
- [x] For who it's designed for
- [x] Explanations or a Link to know how it's work (probably a link to docs)
- [x] Link for contact
- [x] How to contribute
|
1.0
|
Update Readme - We need to update [Readme.md] to keep informations up to date.
- [x] Quick presentation for the project
- [x] Explanation of the goal of the project
- [x] For who it's designed for
- [x] Explanations or a Link to know how it's work (probably a link to docs)
- [x] Link for contact
- [x] How to contribute
|
non_process
|
update readme we need to update to keep informations up to date quick presentation for the project explanation of the goal of the project for who it s designed for explanations or a link to know how it s work probably a link to docs link for contact how to contribute
| 0
|
287,577
| 21,660,642,673
|
IssuesEvent
|
2022-05-06 18:40:02
|
dpaceoffice/ECommerce
|
https://api.github.com/repos/dpaceoffice/ECommerce
|
opened
|
Server Design
|
documentation
|
#7
Create a design document that fully defines the app's backend services. This document should
include: complete REST API, identify all server-side logic & data processing responsibilities,
define all backend tooling and libraries.
|
1.0
|
Server Design - #7
Create a design document that fully defines the app's backend services. This document should
include: complete REST API, identify all server-side logic & data processing responsibilities,
define all backend tooling and libraries.
|
non_process
|
server design create a design document that fully defines the app s backend services this document should include complete rest api identify all server side logic data processing responsibilities define all backend tooling and libraries
| 0
|
13,882
| 16,654,734,355
|
IssuesEvent
|
2021-06-05 10:06:11
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[PM] Responsive issue in Sites tab > UI issues
|
Bug P2 Participant manager Process: Fixed Process: Tested dev
|
Responsive issue in Sites tab > UI issues
1. Add site button is wrapping up with the above text

2. There should be a space between images and text and other UI issues

|
2.0
|
[PM] Responsive issue in Sites tab > UI issues - Responsive issue in Sites tab > UI issues
1. Add site button is wrapping up with the above text

2. There should be a space between images and text and other UI issues

|
process
|
responsive issue in sites tab ui issues responsive issue in sites tab ui issues add site button is wrapping up with the above text there should be a space between images and text and other ui issues
| 1
|
153,864
| 13,529,051,284
|
IssuesEvent
|
2020-09-15 17:40:16
|
aws/chalice
|
https://api.github.com/repos/aws/chalice
|
closed
|
chalice deploy import error, with layers
|
documentation
|
Hi
I got module not found error when deploying, I specify the function to use layers
`{
"version": "2.0",
"app_name": "myapp",
"stages": {
"dev": {
"api_gateway_stage": "api",
"manage_iam_role": false,
"iam_role_arn": "arn",
"lambda_functions": {
"myfunc": {
"layers": ["layer1", "layer2"]
}
}
}
},
}`
I did not specify the module in requirements.txt since I have the layers specified. I probably missing something, could not find much docs on layers for chalice.
Thanks
Dong
|
1.0
|
chalice deploy import error, with layers - Hi
I got module not found error when deploying, I specify the function to use layers
`{
"version": "2.0",
"app_name": "myapp",
"stages": {
"dev": {
"api_gateway_stage": "api",
"manage_iam_role": false,
"iam_role_arn": "arn",
"lambda_functions": {
"myfunc": {
"layers": ["layer1", "layer2"]
}
}
}
},
}`
I did not specify the module in requirements.txt since I have the layers specified. I probably missing something, could not find much docs on layers for chalice.
Thanks
Dong
|
non_process
|
chalice deploy import error with layers hi i got module not found error when deploying i specify the function to use layers version app name myapp stages dev api gateway stage api manage iam role false iam role arn arn lambda functions myfunc layers i did not specify the module in requirements txt since i have the layers specified i probably missing something could not find much docs on layers for chalice thanks dong
| 0
|
97,350
| 28,209,870,970
|
IssuesEvent
|
2023-04-05 02:36:13
|
opensearch-project/performance-analyzer
|
https://api.github.com/repos/opensearch-project/performance-analyzer
|
closed
|
performance-analyzer.log is not rotated
|
bug Build Libraries & Interfaces
|
**Describe the bug**
We are running a three node cluster on Docker and our monitoring system reported that the node disks were running out of space. After digging out we have noticed that `/usr/share/opensearch/logs/performance-analyzer.log` was not rotated (the file was huge, several GiB).
I went through the documentation and found nothing on the subject.
I have to run this command once in a while on each Opensearch node to workaround the issue:
```
docker exec <opensearch_container> truncate -s0 /usr/share/opensearch/logs/performance-analyzer.log
```
**To Reproduce**
Steps to reproduce the behavior:
1. Deploy an Opensearch on Docker
2. Let it run for a while
**Expected behavior**
`/usr/share/opensearch/logs/performance-analyzer.log` should be rotated
**Plugins**
Security plugin
**Host/Environment:**
- OS: Debian Bullseye
- Opensearch Docker image opensearchproject/opensearch:1.2.4
- Docker CE 20.10.12
|
1.0
|
performance-analyzer.log is not rotated - **Describe the bug**
We are running a three node cluster on Docker and our monitoring system reported that the node disks were running out of space. After digging out we have noticed that `/usr/share/opensearch/logs/performance-analyzer.log` was not rotated (the file was huge, several GiB).
I went through the documentation and found nothing on the subject.
I have to run this command once in a while on each Opensearch node to workaround the issue:
```
docker exec <opensearch_container> truncate -s0 /usr/share/opensearch/logs/performance-analyzer.log
```
**To Reproduce**
Steps to reproduce the behavior:
1. Deploy an Opensearch on Docker
2. Let it run for a while
**Expected behavior**
`/usr/share/opensearch/logs/performance-analyzer.log` should be rotated
**Plugins**
Security plugin
**Host/Environment:**
- OS: Debian Bullseye
- Opensearch Docker image opensearchproject/opensearch:1.2.4
- Docker CE 20.10.12
|
non_process
|
performance analyzer log is not rotated describe the bug we are running a three node cluster on docker and our monitoring system reported that the node disks were running out of space after digging out we have noticed that usr share opensearch logs performance analyzer log was not rotated the file was huge several gib i went through the documentation and found nothing on the subject i have to run this command once in a while on each opensearch node to workaround the issue docker exec truncate usr share opensearch logs performance analyzer log to reproduce steps to reproduce the behavior deploy an opensearch on docker let it run for a while expected behavior usr share opensearch logs performance analyzer log should be rotated plugins security plugin host environment os debian bullseye opensearch docker image opensearchproject opensearch docker ce
| 0
|
6,216
| 9,126,231,862
|
IssuesEvent
|
2019-02-24 19:59:41
|
rtcharity/eahub.org
|
https://api.github.com/repos/rtcharity/eahub.org
|
opened
|
Periodic check-ins?
|
Process
|
I'm of two minds as to whether we should restart doing periodic check-ins of what people are working on. On the one hand, it's useful for me to know what everyone else is doing so that I can factor that into what I'm doing, e.g., avoid duplicating work. And I'm sure the product and project managers would like to know this too. On the other hand they mean committing to deadlines for short-term tasks much more frequently (since people are counting on you doing what you said you'd do), which I'm a bit reluctant to do since I'm doing all of this in what spare time I can manage. And I don't want to wind up flaking again. What do other people think?
|
1.0
|
Periodic check-ins? - I'm of two minds as to whether we should restart doing periodic check-ins of what people are working on. On the one hand, it's useful for me to know what everyone else is doing so that I can factor that into what I'm doing, e.g., avoid duplicating work. And I'm sure the product and project managers would like to know this too. On the other hand they mean committing to deadlines for short-term tasks much more frequently (since people are counting on you doing what you said you'd do), which I'm a bit reluctant to do since I'm doing all of this in what spare time I can manage. And I don't want to wind up flaking again. What do other people think?
|
process
|
periodic check ins i m of two minds as to whether we should restart doing periodic check ins of what people are working on on the one hand it s useful for me to know what everyone else is doing so that i can factor that into what i m doing e g avoid duplicating work and i m sure the product and project managers would like to know this too on the other hand they mean committing to deadlines for short term tasks much more frequently since people are counting on you doing what you said you d do which i m a bit reluctant to do since i m doing all of this in what spare time i can manage and i don t want to wind up flaking again what do other people think
| 1
|
20,594
| 27,263,788,878
|
IssuesEvent
|
2023-02-22 16:33:17
|
MicrosoftDocs/azure-devops-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
|
closed
|
Why can't we reference template files from an artifact?
|
devops/prod doc-bug Pri1 devops-cicd-process/tech
|
This section of the document:
https://docs.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops#insert-a-template
States that "Template files need to exist on your filesystem at the start of a pipeline run. You can't reference templates in an artifact."
If we want to reuse a common repo for release template steps, it is important that we use a snapshot version of the templates steps at the time of build. That way the release steps are known and predictable throughout all stages of a release.
Referencing release steps based on a repo filesystem path is dangerous because changes to that template for an in-flight release could break or introduce unwanted changes.
Can you please clarify if there is any workaround to reference release templates from the current build artifact?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 6724abea-bbdc-bf66-ed5e-3214fa6c3e66
* Version Independent ID: 4f8dab21-3f0e-da32-cc0e-1d85c13c0065
* Content: [Templates - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops#step-re-use)
* Content Source: [docs/pipelines/process/templates.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/templates.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
1.0
|
Why can't we reference template files from an artifact? - This section of the document:
https://docs.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops#insert-a-template
States that "Template files need to exist on your filesystem at the start of a pipeline run. You can't reference templates in an artifact."
If we want to reuse a common repo for release template steps, it is important that we use a snapshot version of the templates steps at the time of build. That way the release steps are known and predictable throughout all stages of a release.
Referencing release steps based on a repo filesystem path is dangerous because changes to that template for an in-flight release could break or introduce unwanted changes.
Can you please clarify if there is any workaround to reference release templates from the current build artifact?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 6724abea-bbdc-bf66-ed5e-3214fa6c3e66
* Version Independent ID: 4f8dab21-3f0e-da32-cc0e-1d85c13c0065
* Content: [Templates - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops#step-re-use)
* Content Source: [docs/pipelines/process/templates.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/templates.md)
* Product: **devops**
* Technology: **devops-cicd-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam**
|
process
|
why can t we reference template files from an artifact this section of the document states that template files need to exist on your filesystem at the start of a pipeline run you can t reference templates in an artifact if we want to reuse a common repo for release template steps it is important that we use a snapshot version of the templates steps at the time of build that way the release steps are known and predictable throughout all stages of a release referencing release steps based on a repo filesystem path is dangerous because changes to that template for an in flight release could break or introduce unwanted changes can you please clarify if there is any workaround to reference release templates from the current build artifact document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id bbdc version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
| 1
|
79,286
| 28,086,167,588
|
IssuesEvent
|
2023-03-30 09:54:57
|
snowplow/snowplow-javascript-tracker
|
https://api.github.com/repos/snowplow/snowplow-javascript-tracker
|
closed
|
Fix top level domain grouping initial cookie set attempt
|
type:defect
|
**Describe the bug**
The algorithm behind the `discoverRootDomain` functionality checks if the domain is valid to set a cookie at the highest level domain. This algorithm currently starts from a eTLD which will not be valid in any case, e.g. `.com`, `.io` etc.
This behaviour will create a console warning every time in some browsers e.g. Firefox

The fix would just be to start trying from the TLD and not the eTLD.
**Note**
Even with the fix, if we have an eTLD with more than one part e.g. co.uk, then the warning will still be there. This happens with other trackers as well e.g. analytics.js
|
1.0
|
Fix top level domain grouping initial cookie set attempt - **Describe the bug**
The algorithm behind the `discoverRootDomain` functionality checks if the domain is valid to set a cookie at the highest level domain. This algorithm currently starts from a eTLD which will not be valid in any case, e.g. `.com`, `.io` etc.
This behaviour will create a console warning every time in some browsers e.g. Firefox

The fix would just be to start trying from the TLD and not the eTLD.
**Note**
Even with the fix, if we have an eTLD with more than one part e.g. co.uk, then the warning will still be there. This happens with other trackers as well e.g. analytics.js
|
non_process
|
fix top level domain grouping initial cookie set attempt describe the bug the algorithm behind the discoverrootdomain functionality checks if the domain is valid to set a cookie at the highest level domain this algorithm currently starts from a etld which will not be valid in any case e g com io etc this behaviour will create a console warning every time in some browsers e g firefox the fix would just be to start trying from the tld and not the etld note even with the fix if we have an etld with more than one part e g co uk then the warning will still be there this happens with other trackers as well e g analytics js
| 0
|
14,329
| 17,362,631,044
|
IssuesEvent
|
2021-07-29 23:43:48
|
googleapis/python-spanner
|
https://api.github.com/repos/googleapis/python-spanner
|
closed
|
tests.system.test_system.TestSessionAPI: many tests failed
|
api: spanner flakybot: issue type: process
|
Many tests failed at the same time in this package.
* I will close this issue when there are no more failures in this package _and_
there is at least one pass.
* No new issues will be filed for this package until this issue is closed.
* If there are already issues for individual test cases, I will close them when
the corresponding test passes. You can close them earlier, if you prefer, and
I won't reopen them while this issue is still open.
Here are the tests that failed:
* test_batch_insert_or_update_then_query
* test_batch_insert_then_read
* test_batch_insert_then_read_all_datatypes
* test_batch_insert_then_read_string_array_of_string
* test_batch_insert_w_commit_timestamp
* test_empty_read
* test_empty_read_with_single_keys_index
* test_execute_partitioned_dml
* test_execute_sql_returning_array_of_struct
* test_execute_sql_returning_empty_array_of_struct
* test_execute_sql_returning_transfinite_floats
* test_execute_sql_select_1
* test_execute_sql_w_bool_bindings
* test_execute_sql_w_bytes_bindings
* test_execute_sql_w_date_bindings
* test_execute_sql_w_float64_bindings
* test_execute_sql_w_float_bindings_transfinite
* test_execute_sql_w_int64_bindings
* test_execute_sql_w_manual_consume
* test_execute_sql_w_numeric_bindings
* test_execute_sql_w_query_param_struct
* test_execute_sql_w_string_bindings
* test_execute_sql_w_timestamp_bindings
* test_invalid_type
* test_multiuse_snapshot_execute_sql_isolation_strong
* test_multiuse_snapshot_read_isolation_exact_staleness
* test_multiuse_snapshot_read_isolation_read_timestamp
* test_multiuse_snapshot_read_isolation_strong
* test_partition_query
* test_partition_read_w_index
* test_read_partial_range_from_beginning
* test_read_partial_range_until_end
* test_read_w_index
* test_read_w_limit
* test_read_w_multiple_keys
* test_read_w_ranges
* test_read_w_single_key
* test_read_with_multiple_keys_index
* test_read_with_range_keys_and_index_closed_closed
* test_read_with_range_keys_and_index_closed_open
* test_read_with_range_keys_and_index_open_closed
* test_read_with_range_keys_and_index_open_open
* test_read_with_range_keys_index_closed_closed
* test_read_with_range_keys_index_closed_open
* test_read_with_range_keys_index_limit_closed_closed
* test_read_with_range_keys_index_limit_closed_open
* test_read_with_range_keys_index_limit_open_closed
* test_read_with_range_keys_index_limit_open_open
* test_read_with_range_keys_index_open_closed
* test_read_with_range_keys_index_open_open
* test_read_with_range_keys_index_single_key
* test_read_with_single_keys_index
* test_session_crud
* test_snapshot_read_w_various_staleness
* test_transaction_batch_update_and_execute_dml
* test_transaction_batch_update_success
* test_transaction_batch_update_w_parent_span
* test_transaction_batch_update_w_syntax_error
* test_transaction_batch_update_wo_statements
* test_transaction_execute_sql_w_dml_read_rollback
* test_transaction_execute_update_read_commit
* test_transaction_execute_update_then_insert_commit
* test_transaction_query_w_concurrent_updates
* test_transaction_read_and_insert_or_update_then_commit
* test_transaction_read_and_insert_then_exception
* test_transaction_read_and_insert_then_rollback
* test_transaction_read_w_abort
* test_transaction_read_w_concurrent_updates
-----
commit: 2487800e31842a44dcc37937c325e130c8c926b0
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/306a2e02-87cb-4be9-be31-37456ec7a8a2), [Sponge](http://sponge2/306a2e02-87cb-4be9-be31-37456ec7a8a2)
status: failed
|
1.0
|
tests.system.test_system.TestSessionAPI: many tests failed - Many tests failed at the same time in this package.
* I will close this issue when there are no more failures in this package _and_
there is at least one pass.
* No new issues will be filed for this package until this issue is closed.
* If there are already issues for individual test cases, I will close them when
the corresponding test passes. You can close them earlier, if you prefer, and
I won't reopen them while this issue is still open.
Here are the tests that failed:
* test_batch_insert_or_update_then_query
* test_batch_insert_then_read
* test_batch_insert_then_read_all_datatypes
* test_batch_insert_then_read_string_array_of_string
* test_batch_insert_w_commit_timestamp
* test_empty_read
* test_empty_read_with_single_keys_index
* test_execute_partitioned_dml
* test_execute_sql_returning_array_of_struct
* test_execute_sql_returning_empty_array_of_struct
* test_execute_sql_returning_transfinite_floats
* test_execute_sql_select_1
* test_execute_sql_w_bool_bindings
* test_execute_sql_w_bytes_bindings
* test_execute_sql_w_date_bindings
* test_execute_sql_w_float64_bindings
* test_execute_sql_w_float_bindings_transfinite
* test_execute_sql_w_int64_bindings
* test_execute_sql_w_manual_consume
* test_execute_sql_w_numeric_bindings
* test_execute_sql_w_query_param_struct
* test_execute_sql_w_string_bindings
* test_execute_sql_w_timestamp_bindings
* test_invalid_type
* test_multiuse_snapshot_execute_sql_isolation_strong
* test_multiuse_snapshot_read_isolation_exact_staleness
* test_multiuse_snapshot_read_isolation_read_timestamp
* test_multiuse_snapshot_read_isolation_strong
* test_partition_query
* test_partition_read_w_index
* test_read_partial_range_from_beginning
* test_read_partial_range_until_end
* test_read_w_index
* test_read_w_limit
* test_read_w_multiple_keys
* test_read_w_ranges
* test_read_w_single_key
* test_read_with_multiple_keys_index
* test_read_with_range_keys_and_index_closed_closed
* test_read_with_range_keys_and_index_closed_open
* test_read_with_range_keys_and_index_open_closed
* test_read_with_range_keys_and_index_open_open
* test_read_with_range_keys_index_closed_closed
* test_read_with_range_keys_index_closed_open
* test_read_with_range_keys_index_limit_closed_closed
* test_read_with_range_keys_index_limit_closed_open
* test_read_with_range_keys_index_limit_open_closed
* test_read_with_range_keys_index_limit_open_open
* test_read_with_range_keys_index_open_closed
* test_read_with_range_keys_index_open_open
* test_read_with_range_keys_index_single_key
* test_read_with_single_keys_index
* test_session_crud
* test_snapshot_read_w_various_staleness
* test_transaction_batch_update_and_execute_dml
* test_transaction_batch_update_success
* test_transaction_batch_update_w_parent_span
* test_transaction_batch_update_w_syntax_error
* test_transaction_batch_update_wo_statements
* test_transaction_execute_sql_w_dml_read_rollback
* test_transaction_execute_update_read_commit
* test_transaction_execute_update_then_insert_commit
* test_transaction_query_w_concurrent_updates
* test_transaction_read_and_insert_or_update_then_commit
* test_transaction_read_and_insert_then_exception
* test_transaction_read_and_insert_then_rollback
* test_transaction_read_w_abort
* test_transaction_read_w_concurrent_updates
-----
commit: 2487800e31842a44dcc37937c325e130c8c926b0
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/306a2e02-87cb-4be9-be31-37456ec7a8a2), [Sponge](http://sponge2/306a2e02-87cb-4be9-be31-37456ec7a8a2)
status: failed
|
process
|
tests system test system testsessionapi many tests failed many tests failed at the same time in this package i will close this issue when there are no more failures in this package and there is at least one pass no new issues will be filed for this package until this issue is closed if there are already issues for individual test cases i will close them when the corresponding test passes you can close them earlier if you prefer and i won t reopen them while this issue is still open here are the tests that failed test batch insert or update then query test batch insert then read test batch insert then read all datatypes test batch insert then read string array of string test batch insert w commit timestamp test empty read test empty read with single keys index test execute partitioned dml test execute sql returning array of struct test execute sql returning empty array of struct test execute sql returning transfinite floats test execute sql select test execute sql w bool bindings test execute sql w bytes bindings test execute sql w date bindings test execute sql w bindings test execute sql w float bindings transfinite test execute sql w bindings test execute sql w manual consume test execute sql w numeric bindings test execute sql w query param struct test execute sql w string bindings test execute sql w timestamp bindings test invalid type test multiuse snapshot execute sql isolation strong test multiuse snapshot read isolation exact staleness test multiuse snapshot read isolation read timestamp test multiuse snapshot read isolation strong test partition query test partition read w index test read partial range from beginning test read partial range until end test read w index test read w limit test read w multiple keys test read w ranges test read w single key test read with multiple keys index test read with range keys and index closed closed test read with range keys and index closed open test read with range keys and index open closed test read with range keys and index open open test read with range keys index closed closed test read with range keys index closed open test read with range keys index limit closed closed test read with range keys index limit closed open test read with range keys index limit open closed test read with range keys index limit open open test read with range keys index open closed test read with range keys index open open test read with range keys index single key test read with single keys index test session crud test snapshot read w various staleness test transaction batch update and execute dml test transaction batch update success test transaction batch update w parent span test transaction batch update w syntax error test transaction batch update wo statements test transaction execute sql w dml read rollback test transaction execute update read commit test transaction execute update then insert commit test transaction query w concurrent updates test transaction read and insert or update then commit test transaction read and insert then exception test transaction read and insert then rollback test transaction read w abort test transaction read w concurrent updates commit buildurl status failed
| 1
|
79,015
| 7,692,912,230
|
IssuesEvent
|
2018-05-18 00:13:55
|
tnibert/project-camelot
|
https://api.github.com/repos/tnibert/project-camelot
|
closed
|
Make username login case insensitive
|
expandtests
|
- Change username in user login and registration to convert to lower case
- Go through existing db and verify case, change if necessary
|
1.0
|
Make username login case insensitive - - Change username in user login and registration to convert to lower case
- Go through existing db and verify case, change if necessary
|
non_process
|
make username login case insensitive change username in user login and registration to convert to lower case go through existing db and verify case change if necessary
| 0
|
586,047
| 17,568,669,458
|
IssuesEvent
|
2021-08-14 07:56:44
|
xnd-project/rbc
|
https://api.github.com/repos/xnd-project/rbc
|
opened
|
rbc 0.6.0 release plan
|
high-priority
|
For rbc 0.6.0, the following issues/PRs must be resolved/landed:
- [ ] https://github.com/omnisci/omniscidb-internal/pull/5765 - required to support `set_output_row_size` within rbc
- [ ] https://github.com/xnd-project/rbc/pull/341 - implements `set_output_row_size` support in rbc
- [ ] Optionally https://github.com/xnd-project/rbc/pull/353 - implements numba 0.54 support
|
1.0
|
rbc 0.6.0 release plan - For rbc 0.6.0, the following issues/PRs must be resolved/landed:
- [ ] https://github.com/omnisci/omniscidb-internal/pull/5765 - required to support `set_output_row_size` within rbc
- [ ] https://github.com/xnd-project/rbc/pull/341 - implements `set_output_row_size` support in rbc
- [ ] Optionally https://github.com/xnd-project/rbc/pull/353 - implements numba 0.54 support
|
non_process
|
rbc release plan for rbc the following issues prs must be resolved landed required to support set output row size within rbc implements set output row size support in rbc optionally implements numba support
| 0
|
13,696
| 16,454,792,657
|
IssuesEvent
|
2021-05-21 10:59:28
|
ORNL-AMO/AMO-Tools-Desktop
|
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
|
opened
|
Adding PHAST material updates options in modification
|
Process Heating
|
In some process heating calculators adding a new material/gas/solid should also update the options for the modification side. check PHAST modify-conditions also.
|
1.0
|
Adding PHAST material updates options in modification - In some process heating calculators adding a new material/gas/solid should also update the options for the modification side. check PHAST modify-conditions also.
|
process
|
adding phast material updates options in modification in some process heating calculators adding a new material gas solid should also update the options for the modification side check phast modify conditions also
| 1
|
308,552
| 26,611,490,856
|
IssuesEvent
|
2023-01-24 00:49:06
|
OvercastCommunity/CommunityMaps
|
https://api.github.com/repos/OvercastCommunity/CommunityMaps
|
opened
|
[Payload Race/10v10] TF2: Pipeline
|
map submission contest
|
### Checklist
- [X] I have [pruned](https://pgm.dev/docs/guides/packaging/pruning-chunks) the map.
- [X] I have agreed with assigning the CC BY-SA license to this map, as mentioned in the [README](https://github.com/OvercastCommunity/CommunityMaps#readme).
- [X] I have created an XML file.
- [X] I have created a map image.
- [X] I have uploaded the map zip file to a file sharing service.
- [X] The map has been tested locally to make sure it works.
### Map Name
TF2: Pipeline
### Gamemode & Map Description
_TF2: Pipeline_ is a highly experimental Conquest map, using an alternate form of the new Payload gamemode. Instead of using Attack/Defense, this map is a Payload Race wherein each team has their own payload, and much attempt to attack and defend at the same time. As a proof of concept, this is an adaptation of the first stage of the map Pipeline from Team Fortress 2, with some creative liberties with gameplay and aesthetics in mind. The map has some cover for bow fire and also multiple paths to encourage melee flanking as well. With the current xml setup, it's possible to stop the enemy team's cart movement by entering its range, carts take 2 minutes of nonstop movement to hit the enemy end zone, and carts will slowly lose progress with no teammates within its range. There are alternates to all of these, so they can be changed in the future beyond the contest if necessary.
### Team Sizes
10v10
### Screenshots
https://imgur.com/a/SNoaSrq
### XML
https://gist.github.com/Antnecb/2d45436b14d67f531ab6406a0c20e217
### Map Image

### Map Download
https://drive.google.com/file/d/1UOyvwSJI7MoCbwp8OmmLjApInBq-V_BX/view?usp=sharing
|
1.0
|
[Payload Race/10v10] TF2: Pipeline - ### Checklist
- [X] I have [pruned](https://pgm.dev/docs/guides/packaging/pruning-chunks) the map.
- [X] I have agreed with assigning the CC BY-SA license to this map, as mentioned in the [README](https://github.com/OvercastCommunity/CommunityMaps#readme).
- [X] I have created an XML file.
- [X] I have created a map image.
- [X] I have uploaded the map zip file to a file sharing service.
- [X] The map has been tested locally to make sure it works.
### Map Name
TF2: Pipeline
### Gamemode & Map Description
_TF2: Pipeline_ is a highly experimental Conquest map, using an alternate form of the new Payload gamemode. Instead of using Attack/Defense, this map is a Payload Race wherein each team has their own payload, and much attempt to attack and defend at the same time. As a proof of concept, this is an adaptation of the first stage of the map Pipeline from Team Fortress 2, with some creative liberties with gameplay and aesthetics in mind. The map has some cover for bow fire and also multiple paths to encourage melee flanking as well. With the current xml setup, it's possible to stop the enemy team's cart movement by entering its range, carts take 2 minutes of nonstop movement to hit the enemy end zone, and carts will slowly lose progress with no teammates within its range. There are alternates to all of these, so they can be changed in the future beyond the contest if necessary.
### Team Sizes
10v10
### Screenshots
https://imgur.com/a/SNoaSrq
### XML
https://gist.github.com/Antnecb/2d45436b14d67f531ab6406a0c20e217
### Map Image

### Map Download
https://drive.google.com/file/d/1UOyvwSJI7MoCbwp8OmmLjApInBq-V_BX/view?usp=sharing
|
non_process
|
pipeline checklist i have the map i have agreed with assigning the cc by sa license to this map as mentioned in the i have created an xml file i have created a map image i have uploaded the map zip file to a file sharing service the map has been tested locally to make sure it works map name pipeline gamemode map description pipeline is a highly experimental conquest map using an alternate form of the new payload gamemode instead of using attack defense this map is a payload race wherein each team has their own payload and much attempt to attack and defend at the same time as a proof of concept this is an adaptation of the first stage of the map pipeline from team fortress with some creative liberties with gameplay and aesthetics in mind the map has some cover for bow fire and also multiple paths to encourage melee flanking as well with the current xml setup it s possible to stop the enemy team s cart movement by entering its range carts take minutes of nonstop movement to hit the enemy end zone and carts will slowly lose progress with no teammates within its range there are alternates to all of these so they can be changed in the future beyond the contest if necessary team sizes screenshots xml map image map download
| 0
|
1,058
| 3,526,135,905
|
IssuesEvent
|
2016-01-14 02:22:20
|
osresearch/vst
|
https://api.github.com/repos/osresearch/vst
|
opened
|
VstBuffer should track segments
|
processing
|
The `VstBuffer` class should track segments, rather than individual points. This might make the sort algorithm simpler and maps well to the `line()` override.
|
1.0
|
VstBuffer should track segments - The `VstBuffer` class should track segments, rather than individual points. This might make the sort algorithm simpler and maps well to the `line()` override.
|
process
|
vstbuffer should track segments the vstbuffer class should track segments rather than individual points this might make the sort algorithm simpler and maps well to the line override
| 1
|
22,367
| 31,116,577,417
|
IssuesEvent
|
2023-08-15 00:35:32
|
hsmusic/hsmusic-wiki
|
https://api.github.com/repos/hsmusic/hsmusic-wiki
|
opened
|
Remove "Has Track Art" field from albums & ensure track covers are always associated with credits
|
scope: data processing scope: data format - YAML scope: data format - thing objects
|
Direct quote from Discord #code-quarantine as it covers the situation sufficiently:
---
```js
// Now this is a doozy!
static hasUniqueCoverArt(
track,
albumData,
coverArtistContribsByRef,
hasCoverArt
) {
if (coverArtistContribsByRef?.length > 0) {
return true;
} else if (coverArtistContribsByRef) {
return false;
} else if (hasCoverArt === false) {
return false;
} else if (Track.findAlbum(track, albumData)?.hasTrackArt) {
return true;
} else {
return false;
}
}
```
understanding:
1. if `coverArtistContribsByRef` is non-null (`?.`) and contains at least one item (`.length > 0`), return true
2. if `coverArtistContribsByRef` is non-null (truthy) and contains zero items (by implication), return false
3. if `hasCoverArt` is explicitly set to false for this track, return false
4. if hasTrackArt is true for the album - which it is, unless explicitly set to false - return true
5. return false
do not understand why pt 2 exists at all. data files never set a field to an empty array. this is possible (they are normal YAML files so `Cover Artists: []` is valid), but never happens in practice
pt 1 returns true in the correct case, but it is only relevant in the case where the album is `Has Track Art: false`
else, listing no cover artists is tantamount to the wiki giving up and saying "OK, well, here is this track's art, but who knows who made it" ¯\_(ツ)_/¯
proposed simplified logic:
1. if `coverArtistContribsByRef` is non-null and contains at least one item, return true (use `empty()`)
2. if `hasCoverArt` is explicitly set to false for this track, return false
3. if the album's `trackCoverArtistContribsByRef` is non-null and contains at least one item, return true
4. return false
pt 1 is the end-all case. if cover artist credits are present, then the track has cover art. period.
pt 2 exists solely to override pt 3. if pt 3 is irrelevant then `hasCoverArt` should not be set to false, because pt 3 would fall through to pt 4, return false.
pt 3 exists to ensure tracks automatically have unique artwork if a `Default Track Cover Artists` field has been set. (again, this can be overridden by pt 2 if select tracks do not have cover artworks while the rest share the same cover artists.)
this altogether nullifies a need for the `Has Track Art` field (on albums).
also, it guarantees that all track artworks are associated with cover artist credits. both "return true" cases are directly the result of cover artist contributions being present. (either directly on the track, or inherited from the album)
|
1.0
|
Remove "Has Track Art" field from albums & ensure track covers are always associated with credits - Direct quote from Discord #code-quarantine as it covers the situation sufficiently:
---
```js
// Now this is a doozy!
static hasUniqueCoverArt(
track,
albumData,
coverArtistContribsByRef,
hasCoverArt
) {
if (coverArtistContribsByRef?.length > 0) {
return true;
} else if (coverArtistContribsByRef) {
return false;
} else if (hasCoverArt === false) {
return false;
} else if (Track.findAlbum(track, albumData)?.hasTrackArt) {
return true;
} else {
return false;
}
}
```
understanding:
1. if `coverArtistContribsByRef` is non-null (`?.`) and contains at least one item (`.length > 0`), return true
2. if `coverArtistContribsByRef` is non-null (truthy) and contains zero items (by implication), return false
3. if `hasCoverArt` is explicitly set to false for this track, return false
4. if hasTrackArt is true for the album - which it is, unless explicitly set to false - return true
5. return false
do not understand why pt 2 exists at all. data files never set a field to an empty array. this is possible (they are normal YAML files so `Cover Artists: []` is valid), but never happens in practice
pt 1 returns true in the correct case, but it is only relevant in the case where the album is `Has Track Art: false`
else, listing no cover artists is tantamount to the wiki giving up and saying "OK, well, here is this track's art, but who knows who made it" ¯\_(ツ)_/¯
proposed simplified logic:
1. if `coverArtistContribsByRef` is non-null and contains at least one item, return true (use `empty()`)
2. if `hasCoverArt` is explicitly set to false for this track, return false
3. if the album's `trackCoverArtistContribsByRef` is non-null and contains at least one item, return true
4. return false
pt 1 is the end-all case. if cover artist credits are present, then the track has cover art. period.
pt 2 exists solely to override pt 3. if pt 3 is irrelevant then `hasCoverArt` should not be set to false, because pt 3 would fall through to pt 4, return false.
pt 3 exists to ensure tracks automatically have unique artwork if a `Default Track Cover Artists` field has been set. (again, this can be overridden by pt 2 if select tracks do not have cover artworks while the rest share the same cover artists.)
this altogether nullifies a need for the `Has Track Art` field (on albums).
also, it guarantees that all track artworks are associated with cover artist credits. both "return true" cases are directly the result of cover artist contributions being present. (either directly on the track, or inherited from the album)
|
process
|
remove has track art field from albums ensure track covers are always associated with credits direct quote from discord code quarantine as it covers the situation sufficiently js now this is a doozy static hasuniquecoverart track albumdata coverartistcontribsbyref hascoverart if coverartistcontribsbyref length return true else if coverartistcontribsbyref return false else if hascoverart false return false else if track findalbum track albumdata hastrackart return true else return false understanding if coverartistcontribsbyref is non null and contains at least one item length return true if coverartistcontribsbyref is non null truthy and contains zero items by implication return false if hascoverart is explicitly set to false for this track return false if hastrackart is true for the album which it is unless explicitly set to false return true return false do not understand why pt exists at all data files never set a field to an empty array this is possible they are normal yaml files so cover artists is valid but never happens in practice pt returns true in the correct case but it is only relevant in the case where the album is has track art false else listing no cover artists is tantamount to the wiki giving up and saying ok well here is this track s art but who knows who made it ¯ ツ ¯ proposed simplified logic if coverartistcontribsbyref is non null and contains at least one item return true use empty if hascoverart is explicitly set to false for this track return false if the album s trackcoverartistcontribsbyref is non null and contains at least one item return true return false pt is the end all case if cover artist credits are present then the track has cover art period pt exists solely to override pt if pt is irrelevant then hascoverart should not be set to false because pt would fall through to pt return false pt exists to ensure tracks automatically have unique artwork if a default track cover artists field has been set again this can be overridden by pt if select tracks do not have cover artworks while the rest share the same cover artists this altogether nullifies a need for the has track art field on albums also it guarantees that all track artworks are associated with cover artist credits both return true cases are directly the result of cover artist contributions being present either directly on the track or inherited from the album
| 1
|
15,416
| 19,604,716,312
|
IssuesEvent
|
2022-01-06 07:53:55
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
GDAL Buffer Vectors; not possible to provide negative buffer distance
|
Processing Bug
|
### What is the bug or the crash?
The `Buffer Distance` field in `Toolbox -> GDAL -> Buffer Vectors` cannot be set to a negative value in the UI. However the tool itself (`ogr2ogr`) IS able to handle negative values properly.
### Steps to reproduce the issue
* Open `Toolbox -> GDAL -> Buffer Vectors`
* Enter negative value in `Buffer Distance` field
* Observe it being set to 0.000
### Versions
QGIS version
3.22.2-Białowieża
QGIS code revision
1601ec46d0
Qt version
5.15.2
Python version
3.9.5
GDAL/OGR version
3.4.0
PROJ version
8.2.0
EPSG Registry database version
v10.038 (2021-10-21)
GEOS version
3.10.0-CAPI-1.16.0
SQLite version
3.35.2
PDAL version
2.3.0
PostgreSQL client version
13.0
SpatiaLite version
5.0.1
QWT version
6.1.3
QScintilla2 version
2.11.5
OS version
Windows 10 Version 2009
Active Python plugins
BGTImport
3.16
inspireNL
2.6
QuickOSM
2.0.0
quick_map_services
0.19.27
db_manager
0.1.20
grassprovider
2.12.99
MetaSearch
0.3.5
processing
2.12.99
sagaprovider
2.12.99
### Supported QGIS version
- [ ] I'm running a supported QGIS version according to the roadmap.
### New profile
- [ ] I tried with a new QGIS profile
### Additional context
_No response_
|
1.0
|
GDAL Buffer Vectors; not possible to provide negative buffer distance - ### What is the bug or the crash?
The `Buffer Distance` field in `Toolbox -> GDAL -> Buffer Vectors` cannot be set to a negative value in the UI. However the tool itself (`ogr2ogr`) IS able to handle negative values properly.
### Steps to reproduce the issue
* Open `Toolbox -> GDAL -> Buffer Vectors`
* Enter negative value in `Buffer Distance` field
* Observe it being set to 0.000
### Versions
QGIS version
3.22.2-Białowieża
QGIS code revision
1601ec46d0
Qt version
5.15.2
Python version
3.9.5
GDAL/OGR version
3.4.0
PROJ version
8.2.0
EPSG Registry database version
v10.038 (2021-10-21)
GEOS version
3.10.0-CAPI-1.16.0
SQLite version
3.35.2
PDAL version
2.3.0
PostgreSQL client version
13.0
SpatiaLite version
5.0.1
QWT version
6.1.3
QScintilla2 version
2.11.5
OS version
Windows 10 Version 2009
Active Python plugins
BGTImport
3.16
inspireNL
2.6
QuickOSM
2.0.0
quick_map_services
0.19.27
db_manager
0.1.20
grassprovider
2.12.99
MetaSearch
0.3.5
processing
2.12.99
sagaprovider
2.12.99
### Supported QGIS version
- [ ] I'm running a supported QGIS version according to the roadmap.
### New profile
- [ ] I tried with a new QGIS profile
### Additional context
_No response_
|
process
|
gdal buffer vectors not possible to provide negative buffer distance what is the bug or the crash the buffer distance field in toolbox gdal buffer vectors cannot be set to a negative value in the ui however the tool itself is able to handle negative values properly steps to reproduce the issue open toolbox gdal buffer vectors enter negative value in buffer distance field observe it being set to versions qgis version białowieża qgis code revision qt version python version gdal ogr version proj version epsg registry database version geos version capi sqlite version pdal version postgresql client version spatialite version qwt version version os version windows version active python plugins bgtimport inspirenl quickosm quick map services db manager grassprovider metasearch processing sagaprovider supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context no response
| 1
|
102,289
| 31,885,091,692
|
IssuesEvent
|
2023-09-16 21:15:48
|
moby/moby
|
https://api.github.com/repos/moby/moby
|
closed
|
No way to use both --cache-from and a local store?
|
area/builder area/distribution version/17.03 area/builder/classic-builder
|
Using the new --cache-from is a huge improvement most of the time, but it seems to cause docker to avoid looking in its, for lack of a better term, local cache, meaning that a second build of the same image starts from scratch, or at least from where the remote cache diverges.
https://github.com/docker/docker/blob/master/daemon/cache.go#L12 seems to confirm that this is very much intended.
It would be great to look at the local cache first and only hit the --cache-from images if it doesn't find anything. Or possibly have a way of specifying the local cache as an option, say
`docker build --cache-from . myimage:v1.0`
It would also be great to document this behavior as part of the `--cache-from` docs. Right now it says "Images to consider as cache sources" but it's unclear that that causes the normal cache to be unconsidered.
Thanks!
**Steps to reproduce the issue:**
```
docker pull myimage:v1.0
docker build --cache-from myimage:v1.0 -t myimage:v1.1 .
docker build --cache-from myimage:v1.0 -t myimage:v1.1 .
```
**Describe the results you received:**
The second build starts from the latest step it shares with myimage:v1.0
**Describe the results you expected:**
The second build re-uses the entire cache from the first build.
**Additional information you deem important (e.g. issue happens only occasionally):**
**Output of `docker version`:**
```
Client:
Version: 17.03.1-ce
API version: 1.27
Go version: go1.7.5
Git commit: c6d412e
Built: Fri Mar 24 00:45:26 2017
OS/Arch: linux/amd64
Server:
Version: 17.03.1-ce
API version: 1.27 (minimum version 1.12)
Go version: go1.7.5
Git commit: c6d412e
Built: Fri Mar 24 00:45:26 2017
OS/Arch: linux/amd64
Experimental: false
```
**Output of `docker info`:**
```
Containers: 12
Running: 0
Paused: 0
Stopped: 12
Images: 239
Server Version: 17.03.1-ce
Storage Driver: overlay2
Backing Filesystem: extfs
Supports d_type: true
Native Overlay Diff: true
Logging Driver: json-file
Cgroup Driver: cgroupfs
Plugins:
Volume: local
Network: bridge host macvlan null overlay
Swarm: inactive
Runtimes: runc
Default Runtime: runc
Init Binary: docker-init
containerd version: 4ab9917febca54791c5f071a9d1f404867857fcc
runc version: 54296cf40ad8143b62dbcaa1d90e520a2136ddfe
init version: 949e6fa
Security Options:
apparmor
seccomp
Profile: default
Kernel Version: 4.4.0-72-generic
Operating System: Ubuntu 16.04.2 LTS
OSType: linux
Architecture: x86_64
CPUs: 2
Total Memory: 7.795 GiB
Name: ip-172-31-50-48
ID: H5BG:BAQV:VVQ2:BGUO:5MQG:OXKU:T6MH:ZDXF:ZMXT:TP7G:B6K3:JIZG
Docker Root Dir: /var/lib/docker
Debug Mode (client): false
Debug Mode (server): false
Username: <redacted>
Registry: https://index.docker.io/v1/
WARNING: No swap limit support
Experimental: false
Insecure Registries:
127.0.0.0/8
Live Restore Enabled: false
```
**Additional environment details (AWS, VirtualBox, physical, etc.):**
|
3.0
|
No way to use both --cache-from and a local store? - Using the new --cache-from is a huge improvement most of the time, but it seems to cause docker to avoid looking in its, for lack of a better term, local cache, meaning that a second build of the same image starts from scratch, or at least from where the remote cache diverges.
https://github.com/docker/docker/blob/master/daemon/cache.go#L12 seems to confirm that this is very much intended.
It would be great to look at the local cache first and only hit the --cache-from images if it doesn't find anything. Or possibly have a way of specifying the local cache as an option, say
`docker build --cache-from . myimage:v1.0`
It would also be great to document this behavior as part of the `--cache-from` docs. Right now it says "Images to consider as cache sources" but it's unclear that that causes the normal cache to be unconsidered.
Thanks!
**Steps to reproduce the issue:**
```
docker pull myimage:v1.0
docker build --cache-from myimage:v1.0 -t myimage:v1.1 .
docker build --cache-from myimage:v1.0 -t myimage:v1.1 .
```
**Describe the results you received:**
The second build starts from the latest step it shares with myimage:v1.0
**Describe the results you expected:**
The second build re-uses the entire cache from the first build.
**Additional information you deem important (e.g. issue happens only occasionally):**
**Output of `docker version`:**
```
Client:
Version: 17.03.1-ce
API version: 1.27
Go version: go1.7.5
Git commit: c6d412e
Built: Fri Mar 24 00:45:26 2017
OS/Arch: linux/amd64
Server:
Version: 17.03.1-ce
API version: 1.27 (minimum version 1.12)
Go version: go1.7.5
Git commit: c6d412e
Built: Fri Mar 24 00:45:26 2017
OS/Arch: linux/amd64
Experimental: false
```
**Output of `docker info`:**
```
Containers: 12
Running: 0
Paused: 0
Stopped: 12
Images: 239
Server Version: 17.03.1-ce
Storage Driver: overlay2
Backing Filesystem: extfs
Supports d_type: true
Native Overlay Diff: true
Logging Driver: json-file
Cgroup Driver: cgroupfs
Plugins:
Volume: local
Network: bridge host macvlan null overlay
Swarm: inactive
Runtimes: runc
Default Runtime: runc
Init Binary: docker-init
containerd version: 4ab9917febca54791c5f071a9d1f404867857fcc
runc version: 54296cf40ad8143b62dbcaa1d90e520a2136ddfe
init version: 949e6fa
Security Options:
apparmor
seccomp
Profile: default
Kernel Version: 4.4.0-72-generic
Operating System: Ubuntu 16.04.2 LTS
OSType: linux
Architecture: x86_64
CPUs: 2
Total Memory: 7.795 GiB
Name: ip-172-31-50-48
ID: H5BG:BAQV:VVQ2:BGUO:5MQG:OXKU:T6MH:ZDXF:ZMXT:TP7G:B6K3:JIZG
Docker Root Dir: /var/lib/docker
Debug Mode (client): false
Debug Mode (server): false
Username: <redacted>
Registry: https://index.docker.io/v1/
WARNING: No swap limit support
Experimental: false
Insecure Registries:
127.0.0.0/8
Live Restore Enabled: false
```
**Additional environment details (AWS, VirtualBox, physical, etc.):**
|
non_process
|
no way to use both cache from and a local store using the new cache from is a huge improvement most of the time but it seems to cause docker to avoid looking in its for lack of a better term local cache meaning that a second build of the same image starts from scratch or at least from where the remote cache diverges seems to confirm that this is very much intended it would be great to look at the local cache first and only hit the cache from images if it doesn t find anything or possibly have a way of specifying the local cache as an option say docker build cache from myimage it would also be great to document this behavior as part of the cache from docs right now it says images to consider as cache sources but it s unclear that that causes the normal cache to be unconsidered thanks steps to reproduce the issue docker pull myimage docker build cache from myimage t myimage docker build cache from myimage t myimage describe the results you received the second build starts from the latest step it shares with myimage describe the results you expected the second build re uses the entire cache from the first build additional information you deem important e g issue happens only occasionally output of docker version client version ce api version go version git commit built fri mar os arch linux server version ce api version minimum version go version git commit built fri mar os arch linux experimental false output of docker info containers running paused stopped images server version ce storage driver backing filesystem extfs supports d type true native overlay diff true logging driver json file cgroup driver cgroupfs plugins volume local network bridge host macvlan null overlay swarm inactive runtimes runc default runtime runc init binary docker init containerd version runc version init version security options apparmor seccomp profile default kernel version generic operating system ubuntu lts ostype linux architecture cpus total memory gib name ip id baqv bguo oxku zdxf zmxt jizg docker root dir var lib docker debug mode client false debug mode server false username registry warning no swap limit support experimental false insecure registries live restore enabled false additional environment details aws virtualbox physical etc
| 0
|
4,325
| 7,236,639,196
|
IssuesEvent
|
2018-02-13 08:04:53
|
muflihun/residue
|
https://api.github.com/repos/muflihun/residue
|
opened
|
Deadlock detected in develop
|
area: log-processing type: bug
|
```
19:03:02,606 [LogHandler] [vTRACE] [session.cc:86] Adding bytes
19:03:02,608 [LogDispatcher] [vDEBUG] [log-request-handler.cc:88] Pausing schedule for client integrity
19:03:02,609 [LogHandler] [vTRACE] [session.cc:75] Received: 522 bytes
19:03:02,609 [LogHandler] [vTRACE] [session.cc:102] Read bytes: 0fe53a0a15ad3b19847a4e5ee30bdcda:muflihun00102030:sHbsBEN2KhHtBETNCpQDU3wHPw27YHBzztBj7yckU2sHE7lLyUw35wCpxM0qALezZcvXC51YAc75cm60NMY1+mJzr4gpHWlSduK4nj8Y87aPa3/GPoBOjpIl1bTYoUghZ6lCyNgBgNmw7I8Av6oxnaiWJ6AsZARX2Oy1K/DMv5kE33XjxEVkLwRSoCGOS71dSo6LtFzF16CzGXqr/bKHB3vBv+xb8Ap9rN2w36dXGk02ydyE5EF18Pt5pLVzEIyhROlG2ErWShBcWdigYy6pY7rl/qlY/9v1GbggQ0n4rUpPjKtS50H+Z4IQ8ZVT63FMEvb6FY6hs654hx7kdssWvA4jwa7r5x0OVr2+lme8qptSpxEJ15eesTYZvgQRXTImMrGXsqo3VXQZ/AQZ6ZSNaW4/99x17SP6SYiVxNIwQgv5WAnh7j+jAO5I7zTMY66kJQtrg7tozUAIBcM8oQed9Q== [size: 522]
19:03:02,609 [LogHandler] [vDEBUG] [session.cc:171] Sending {r:0}
19:03:02,609 [LogHandler] [vTRACE] [session.cc:86] Adding bytes
19:03:02,609 [LogDispatcher] [vCRAZY] [log-request-handler.cc:100] -----============= [ BEGIN ] =============-----
19:03:02,610 [LogDispatcher] [vDEBUG] [request-handler.h:144] Raw request: 938d714753441d7cdf1167023aa7f3f9:muflihun00102030:xLRbv36YW//gqY9gVrsCOBuwoE5NM+A8iJjoE7W1Jq1/PUXOvuCk0lVvlkmGZT2Hq0Vhj3vuzdSWxqiUH2Iu3I5ufo1EbWuIvQb04eZ0k6tZxJ1wRDjhdfIaCUab9Av0nlYohOWi56pD2suVynM4xucvXuFh6mm2T+qeTyF0i7LuZybJ5bouOHlxLzSCTiI+RQNPJqQo/TyCxim5Swa/PNZwD/S9/pPoPY42XDfZehan0r/kDP9kdQw2SPTaBxG79jyqF4g61o0fsnqtXtrVVVtxgbi5ySvSy40TpdYTbG7DME9hfRlGdxciyewftqijfaQePgJweYwTUYxVtl4uttiaPbELbe1OyC6qKvn7drfgFFtIttbUBa8lf8GV8FJ7vmpMr8CZMYk/8V4PYX9kRhiSKgCHK2vUTXlCja7CbYzoTgplL5iwrQf4LNJ/6FdckPS+ZtVHl9knQokhTyGbqw==
19:03:02,610 [LogDispatcher] [vDEBUG] [request-handler.cc:58] Client: muflihun00102030
19:03:02,610 [LogDispatcher] [vCRAZY] [request-handler.cc:59] IV: 938d714753441d7cdf1167023aa7f3f9
19:03:02,610 [LogDispatcher] [vCRAZY] [request-handler.cc:66] Data (base64): xLRbv36YW//gqY9gVrsCOBuwoE5NM+A8iJjoE7W1Jq1/PUXOvuCk0lVvlkmGZT2Hq0Vhj3vuzdSWxqiUH2Iu3I5ufo1EbWuIvQb04eZ0k6tZxJ1wRDjhdfIaCUab9Av0nlYohOWi56pD2suVynM4xucvXuFh6mm2T+qeTyF0i7LuZybJ5bouOHlxLzSCTiI+RQNPJqQo/TyCxim5Swa/PNZwD/S9/pPoPY42XDfZehan0r/kDP9kdQw2SPTaBxG79jyqF4g61o0fsnqtXtrVVVtxgbi5ySvSy40TpdYTbG7DME9hfRlGdxciyewftqijfaQePgJweYwTUYxVtl4uttiaPbELbe1OyC6qKvn7drfgFFtIttbUBa8lf8GV8FJ7vmpMr8CZMYk/8V4PYX9kRhiSKgCHK2vUTXlCja7CbYzoTgplL5iwrQf4LNJ/6FdckPS+ZtVHl9knQokhTyGbqw==
19:03:02,610 [LogDispatcher] [vCRAZY] [request-handler.cc:69] Ripe command: echo 938d714753441d7cdf1167023aa7f3f9:muflihun00102030:xLRbv36YW//gqY9gVrsCOBuwoE5NM+A8iJjoE7W1Jq1/PUXOvuCk0lVvlkmGZT2Hq0Vhj3vuzdSWxqiUH2Iu3I5ufo1EbWuIvQb04eZ0k6tZxJ1wRDjhdfIaCUab9Av0nlYohOWi56pD2suVynM4xucvXuFh6mm2T+qeTyF0i7LuZybJ5bouOHlxLzSCTiI+RQNPJqQo/TyCxim5Swa/PNZwD/S9/pPoPY42XDfZehan0r/kDP9kdQw2SPTaBxG79jyqF4g61o0fsnqtXtrVVVtxgbi5ySvSy40TpdYTbG7DME9hfRlGdxciyewftqijfaQePgJweYwTUYxVtl4uttiaPbELbe1OyC6qKvn7drfgFFtIttbUBa8lf8GV8FJ7vmpMr8CZMYk/8V4PYX9kRhiSKgCHK2vUTXlCja7CbYzoTgplL5iwrQf4LNJ/6FdckPS+ZtVHl9knQokhTyGbqw== | ripe -d --aes --key 00098C6A684633C0087D9C76B47D2A57 --base64
19:03:02,610 [LogDispatcher] [vCRAZY] [request-handler.cc:76] Plain request: eNrtz7FOwzAQxvFXiTxBleA0UdSUmZWtTIjBOEcaNU6D7XRBvDsn9SUY/tvdZ/vOv/cfM7gseQpinvfdvu/q/tg3dV2X5muaNTT2LUlM1n1uYUsufNsXudkoaRo2qfy6Wg3XWZIdJDt9MlQ+uIvYFL116/rkvdFZ2+J11rTkIrhpedCiLPzZxWK3e9TzWW4y6w+aXutp0b1Ne9DyOo4S9eF9R6Xz9HJIo0YnSVmbfI7iBu1fdW5xuncaXy+yaNoejl1bd735LaFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQr1/1M//gDsendT
19:03:02,610 [LogDispatcher] [vTRACE] [request-handler.h:148] Decryption finished (b64): eNrtz7FOwzAQxvFXiTxBleA0UdSUmZWtTIjBOEcaNU6D7XRBvDsn9SUY/tvdZ/vOv/cfM7gseQpinvfdvu/q/tg3dV2X5muaNTT2LUlM1n1uYUsufNsXudkoaRo2qfy6Wg3XWZIdJDt9MlQ+uIvYFL116/rkvdFZ2+J11rTkIrhpedCiLPzZxWK3e9TzWW4y6w+aXutp0b1Ne9DyOo4S9eF9R6Xz9HJIo0YnSVmbfI7iBu1fdW5xuncaXy+yaNoejl1bd735LaFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQr1/1M//gDsendT
19:03:02,610 [LogDispatcher] [vTRACE] [request-handler.h:159] Decompressing: eNrtz7FOwzAQxvFXiTxBleA0UdSUmZWtTIjBOEcaNU6D7XRBvDsn9SUY/tvdZ/vOv/cfM7gseQpinvfdvu/q/tg3dV2X5muaNTT2LUlM1n1uYUsufNsXudkoaRo2qfy6Wg3XWZIdJDt9MlQ+uIvYFL116/rkvdFZ2+J11rTkIrhpedCiLPzZxWK3e9TzWW4y6w+aXutp0b1Ne9DyOo4S9eF9R6Xz9HJIo0YnSVmbfI7iBu1fdW5xuncaXy+yaNoejl1bd735LaFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQr1/1M//gDsendT
19:03:02,611 [LogDispatcher] [vTRACE] [request-handler.h:170] Decompression finished (raw): [{... (removed)}]
19:03:02,612 [LogDispatcher] [vDEBUG] [log-request-handler.cc:115] Is bulk? true
19:03:02,612 [LogDispatcher] [vDEBUG] [log-request-handler.cc:124] Request client: 0x10065cfa8
19:03:02,612 [LogHandler] [vTRACE] [session.cc:75] Received: 522 bytes
19:03:02,612 [LogDispatcher] [vDEBUG] [log-request-handler.cc:206] Force check: true, clientRef: 0x70000f9729a0, *clientRef: muflihun00102030, bypassChecks: false
19:03:02,612 [LogHandler] [vTRACE] [session.cc:102] Read bytes: 29277885c8e3830c20fb91fab830194b:muflihun00102030:zaAqRKANlz9SElDUXg6ePbeykddpAyJjxC3TEL/yutU5HIryqLpN1xmsEncDHizpniIeaCIY7kfUjuzTaXwYJC0dxHxlssq5Lm7PzgPCyyxa7POh0quW1zooTSlBMvUiuCWamAdfzz3WPPjjbMSyvqtgMmbFIhs4qMMMt3kKZbli8kk9Hmv5XV8Ltz+/tLX4ppiOn6qSFh/pIQ0LNaBVK0wHXLfGO+AN58vc1D50JT8vA00XUjkzNNmeasekIntZDoRut/WhgqVijoUsT27sY9WzrsdaLuIJIAUhtewqvAGgyeplI+lucgHWtWKVR+HQQ887giQwW6+80FrU0yIZ/yEhDnFbCul3Zc9L3vuuBHvrAgjYvDPkA4MZPwBkFWSh6anUbJ5zsNgEnMP2tdDI+ZLnSVfvQ+ZNEPeVXYE15la/XVPoe0/b/2MqY3HPUrYEtTJONqT/XrZKl1HXbqxp4w== [size: 522]
19:03:02,613 [LogDispatcher] [vDEBUG] [client.cc:65] Checking token 37953058 (client [muflihun00102030])
```
|
1.0
|
Deadlock detected in develop - ```
19:03:02,606 [LogHandler] [vTRACE] [session.cc:86] Adding bytes
19:03:02,608 [LogDispatcher] [vDEBUG] [log-request-handler.cc:88] Pausing schedule for client integrity
19:03:02,609 [LogHandler] [vTRACE] [session.cc:75] Received: 522 bytes
19:03:02,609 [LogHandler] [vTRACE] [session.cc:102] Read bytes: 0fe53a0a15ad3b19847a4e5ee30bdcda:muflihun00102030:sHbsBEN2KhHtBETNCpQDU3wHPw27YHBzztBj7yckU2sHE7lLyUw35wCpxM0qALezZcvXC51YAc75cm60NMY1+mJzr4gpHWlSduK4nj8Y87aPa3/GPoBOjpIl1bTYoUghZ6lCyNgBgNmw7I8Av6oxnaiWJ6AsZARX2Oy1K/DMv5kE33XjxEVkLwRSoCGOS71dSo6LtFzF16CzGXqr/bKHB3vBv+xb8Ap9rN2w36dXGk02ydyE5EF18Pt5pLVzEIyhROlG2ErWShBcWdigYy6pY7rl/qlY/9v1GbggQ0n4rUpPjKtS50H+Z4IQ8ZVT63FMEvb6FY6hs654hx7kdssWvA4jwa7r5x0OVr2+lme8qptSpxEJ15eesTYZvgQRXTImMrGXsqo3VXQZ/AQZ6ZSNaW4/99x17SP6SYiVxNIwQgv5WAnh7j+jAO5I7zTMY66kJQtrg7tozUAIBcM8oQed9Q== [size: 522]
19:03:02,609 [LogHandler] [vDEBUG] [session.cc:171] Sending {r:0}
19:03:02,609 [LogHandler] [vTRACE] [session.cc:86] Adding bytes
19:03:02,609 [LogDispatcher] [vCRAZY] [log-request-handler.cc:100] -----============= [ BEGIN ] =============-----
19:03:02,610 [LogDispatcher] [vDEBUG] [request-handler.h:144] Raw request: 938d714753441d7cdf1167023aa7f3f9:muflihun00102030:xLRbv36YW//gqY9gVrsCOBuwoE5NM+A8iJjoE7W1Jq1/PUXOvuCk0lVvlkmGZT2Hq0Vhj3vuzdSWxqiUH2Iu3I5ufo1EbWuIvQb04eZ0k6tZxJ1wRDjhdfIaCUab9Av0nlYohOWi56pD2suVynM4xucvXuFh6mm2T+qeTyF0i7LuZybJ5bouOHlxLzSCTiI+RQNPJqQo/TyCxim5Swa/PNZwD/S9/pPoPY42XDfZehan0r/kDP9kdQw2SPTaBxG79jyqF4g61o0fsnqtXtrVVVtxgbi5ySvSy40TpdYTbG7DME9hfRlGdxciyewftqijfaQePgJweYwTUYxVtl4uttiaPbELbe1OyC6qKvn7drfgFFtIttbUBa8lf8GV8FJ7vmpMr8CZMYk/8V4PYX9kRhiSKgCHK2vUTXlCja7CbYzoTgplL5iwrQf4LNJ/6FdckPS+ZtVHl9knQokhTyGbqw==
19:03:02,610 [LogDispatcher] [vDEBUG] [request-handler.cc:58] Client: muflihun00102030
19:03:02,610 [LogDispatcher] [vCRAZY] [request-handler.cc:59] IV: 938d714753441d7cdf1167023aa7f3f9
19:03:02,610 [LogDispatcher] [vCRAZY] [request-handler.cc:66] Data (base64): xLRbv36YW//gqY9gVrsCOBuwoE5NM+A8iJjoE7W1Jq1/PUXOvuCk0lVvlkmGZT2Hq0Vhj3vuzdSWxqiUH2Iu3I5ufo1EbWuIvQb04eZ0k6tZxJ1wRDjhdfIaCUab9Av0nlYohOWi56pD2suVynM4xucvXuFh6mm2T+qeTyF0i7LuZybJ5bouOHlxLzSCTiI+RQNPJqQo/TyCxim5Swa/PNZwD/S9/pPoPY42XDfZehan0r/kDP9kdQw2SPTaBxG79jyqF4g61o0fsnqtXtrVVVtxgbi5ySvSy40TpdYTbG7DME9hfRlGdxciyewftqijfaQePgJweYwTUYxVtl4uttiaPbELbe1OyC6qKvn7drfgFFtIttbUBa8lf8GV8FJ7vmpMr8CZMYk/8V4PYX9kRhiSKgCHK2vUTXlCja7CbYzoTgplL5iwrQf4LNJ/6FdckPS+ZtVHl9knQokhTyGbqw==
19:03:02,610 [LogDispatcher] [vCRAZY] [request-handler.cc:69] Ripe command: echo 938d714753441d7cdf1167023aa7f3f9:muflihun00102030:xLRbv36YW//gqY9gVrsCOBuwoE5NM+A8iJjoE7W1Jq1/PUXOvuCk0lVvlkmGZT2Hq0Vhj3vuzdSWxqiUH2Iu3I5ufo1EbWuIvQb04eZ0k6tZxJ1wRDjhdfIaCUab9Av0nlYohOWi56pD2suVynM4xucvXuFh6mm2T+qeTyF0i7LuZybJ5bouOHlxLzSCTiI+RQNPJqQo/TyCxim5Swa/PNZwD/S9/pPoPY42XDfZehan0r/kDP9kdQw2SPTaBxG79jyqF4g61o0fsnqtXtrVVVtxgbi5ySvSy40TpdYTbG7DME9hfRlGdxciyewftqijfaQePgJweYwTUYxVtl4uttiaPbELbe1OyC6qKvn7drfgFFtIttbUBa8lf8GV8FJ7vmpMr8CZMYk/8V4PYX9kRhiSKgCHK2vUTXlCja7CbYzoTgplL5iwrQf4LNJ/6FdckPS+ZtVHl9knQokhTyGbqw== | ripe -d --aes --key 00098C6A684633C0087D9C76B47D2A57 --base64
19:03:02,610 [LogDispatcher] [vCRAZY] [request-handler.cc:76] Plain request: eNrtz7FOwzAQxvFXiTxBleA0UdSUmZWtTIjBOEcaNU6D7XRBvDsn9SUY/tvdZ/vOv/cfM7gseQpinvfdvu/q/tg3dV2X5muaNTT2LUlM1n1uYUsufNsXudkoaRo2qfy6Wg3XWZIdJDt9MlQ+uIvYFL116/rkvdFZ2+J11rTkIrhpedCiLPzZxWK3e9TzWW4y6w+aXutp0b1Ne9DyOo4S9eF9R6Xz9HJIo0YnSVmbfI7iBu1fdW5xuncaXy+yaNoejl1bd735LaFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQr1/1M//gDsendT
19:03:02,610 [LogDispatcher] [vTRACE] [request-handler.h:148] Decryption finished (b64): eNrtz7FOwzAQxvFXiTxBleA0UdSUmZWtTIjBOEcaNU6D7XRBvDsn9SUY/tvdZ/vOv/cfM7gseQpinvfdvu/q/tg3dV2X5muaNTT2LUlM1n1uYUsufNsXudkoaRo2qfy6Wg3XWZIdJDt9MlQ+uIvYFL116/rkvdFZ2+J11rTkIrhpedCiLPzZxWK3e9TzWW4y6w+aXutp0b1Ne9DyOo4S9eF9R6Xz9HJIo0YnSVmbfI7iBu1fdW5xuncaXy+yaNoejl1bd735LaFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQr1/1M//gDsendT
19:03:02,610 [LogDispatcher] [vTRACE] [request-handler.h:159] Decompressing: eNrtz7FOwzAQxvFXiTxBleA0UdSUmZWtTIjBOEcaNU6D7XRBvDsn9SUY/tvdZ/vOv/cfM7gseQpinvfdvu/q/tg3dV2X5muaNTT2LUlM1n1uYUsufNsXudkoaRo2qfy6Wg3XWZIdJDt9MlQ+uIvYFL116/rkvdFZ2+J11rTkIrhpedCiLPzZxWK3e9TzWW4y6w+aXutp0b1Ne9DyOo4S9eF9R6Xz9HJIo0YnSVmbfI7iBu1fdW5xuncaXy+yaNoejl1bd735LaFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQr1/1M//gDsendT
19:03:02,611 [LogDispatcher] [vTRACE] [request-handler.h:170] Decompression finished (raw): [{... (removed)}]
19:03:02,612 [LogDispatcher] [vDEBUG] [log-request-handler.cc:115] Is bulk? true
19:03:02,612 [LogDispatcher] [vDEBUG] [log-request-handler.cc:124] Request client: 0x10065cfa8
19:03:02,612 [LogHandler] [vTRACE] [session.cc:75] Received: 522 bytes
19:03:02,612 [LogDispatcher] [vDEBUG] [log-request-handler.cc:206] Force check: true, clientRef: 0x70000f9729a0, *clientRef: muflihun00102030, bypassChecks: false
19:03:02,612 [LogHandler] [vTRACE] [session.cc:102] Read bytes: 29277885c8e3830c20fb91fab830194b:muflihun00102030:zaAqRKANlz9SElDUXg6ePbeykddpAyJjxC3TEL/yutU5HIryqLpN1xmsEncDHizpniIeaCIY7kfUjuzTaXwYJC0dxHxlssq5Lm7PzgPCyyxa7POh0quW1zooTSlBMvUiuCWamAdfzz3WPPjjbMSyvqtgMmbFIhs4qMMMt3kKZbli8kk9Hmv5XV8Ltz+/tLX4ppiOn6qSFh/pIQ0LNaBVK0wHXLfGO+AN58vc1D50JT8vA00XUjkzNNmeasekIntZDoRut/WhgqVijoUsT27sY9WzrsdaLuIJIAUhtewqvAGgyeplI+lucgHWtWKVR+HQQ887giQwW6+80FrU0yIZ/yEhDnFbCul3Zc9L3vuuBHvrAgjYvDPkA4MZPwBkFWSh6anUbJ5zsNgEnMP2tdDI+ZLnSVfvQ+ZNEPeVXYE15la/XVPoe0/b/2MqY3HPUrYEtTJONqT/XrZKl1HXbqxp4w== [size: 522]
19:03:02,613 [LogDispatcher] [vDEBUG] [client.cc:65] Checking token 37953058 (client [muflihun00102030])
```
|
process
|
deadlock detected in develop adding bytes pausing schedule for client integrity received bytes read bytes qly sending r adding bytes raw request rqnpjqqo pnzwd client iv data rqnpjqqo pnzwd ripe command echo rqnpjqqo pnzwd ripe d aes key plain request tvdz vov q gdsendt decryption finished tvdz vov q gdsendt decompressing tvdz vov q gdsendt decompression finished raw is bulk true request client received bytes force check true clientref clientref bypasschecks false read bytes lucghwtwkvr zlnsvfvq b checking token client
| 1
|
21,878
| 30,322,790,026
|
IssuesEvent
|
2023-07-10 20:39:04
|
open-telemetry/opentelemetry-collector-contrib
|
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
|
closed
|
Flaky Test - tailsamplingprocessor/TestConcurrentTraceArrival
|
bug help wanted good first issue priority:p3 processor/tailsampling flaky test never stale
|
```
race: limit on 8128 simultaneously alive goroutines is exceeded, dying
FAIL github.com/open-telemetry/opentelemetry-collector-contrib/processor/tailsamplingprocessor 3.415s
```
Observed here: https://github.com/open-telemetry/opentelemetry-collector-contrib/runs/6555994409?check_suite_focus=true
|
1.0
|
Flaky Test - tailsamplingprocessor/TestConcurrentTraceArrival - ```
race: limit on 8128 simultaneously alive goroutines is exceeded, dying
FAIL github.com/open-telemetry/opentelemetry-collector-contrib/processor/tailsamplingprocessor 3.415s
```
Observed here: https://github.com/open-telemetry/opentelemetry-collector-contrib/runs/6555994409?check_suite_focus=true
|
process
|
flaky test tailsamplingprocessor testconcurrenttracearrival race limit on simultaneously alive goroutines is exceeded dying fail github com open telemetry opentelemetry collector contrib processor tailsamplingprocessor observed here
| 1
|
2,293
| 5,114,591,185
|
IssuesEvent
|
2017-01-06 19:00:06
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
closed
|
Multiple aggregations of the same type are missing correct `cols` metadata
|
Bug Priority/P1 Query Processor
|
Similar to #4002, but when two aggregations of the same type are specified, only the first has the correct `cols` metadata, in particular `base_type`, `special_type`, and `source` are incorrect.
Example query:
```json
{
"source_table": 1,
"aggregation": [["count"], ["count"]],
"breakout": [["datetime_field", ["field-id", 1], "as", "day"]]
}
```
Expected `data.cols[1]` and `data.cols[2]`:
```json
{
"description":null,
"table_id":null,
"special_type":"type/Number",
"name":"count",
"source":"aggregation",
"extra_info":{},
"id":null,
"target":null,
"display_name":"count",
"base_type":"type/Integer"
},
{
"description":null,
"table_id":null,
"special_type":"type/Number",
"name":"count_2",
"source":"aggregation",
"extra_info":{},
"id":null,
"target":null,
"display_name":"count_2",
"base_type":"type/Integer"
}
```
Actual `data.cols[1]` and `data.cols[2]`:
```json
{
"description":null,
"table_id":null,
"special_type":"type/Number",
"name":"count",
"source":"aggregation",
"extra_info":{},
"id":null,
"target":null,
"display_name":"count",
"base_type":"type/Integer"
},
{
"description":null,
"table_id":null,
special_type:null,
"name":"count_2",
"extra_info":{},
"id":null,
"target":null,
"preview_display":true,
"display_name":"count_2",
base_type:"type/*"
source:
}
```
|
1.0
|
Multiple aggregations of the same type are missing correct `cols` metadata - Similar to #4002, but when two aggregations of the same type are specified, only the first has the correct `cols` metadata, in particular `base_type`, `special_type`, and `source` are incorrect.
Example query:
```json
{
"source_table": 1,
"aggregation": [["count"], ["count"]],
"breakout": [["datetime_field", ["field-id", 1], "as", "day"]]
}
```
Expected `data.cols[1]` and `data.cols[2]`:
```json
{
"description":null,
"table_id":null,
"special_type":"type/Number",
"name":"count",
"source":"aggregation",
"extra_info":{},
"id":null,
"target":null,
"display_name":"count",
"base_type":"type/Integer"
},
{
"description":null,
"table_id":null,
"special_type":"type/Number",
"name":"count_2",
"source":"aggregation",
"extra_info":{},
"id":null,
"target":null,
"display_name":"count_2",
"base_type":"type/Integer"
}
```
Actual `data.cols[1]` and `data.cols[2]`:
```json
{
"description":null,
"table_id":null,
"special_type":"type/Number",
"name":"count",
"source":"aggregation",
"extra_info":{},
"id":null,
"target":null,
"display_name":"count",
"base_type":"type/Integer"
},
{
"description":null,
"table_id":null,
special_type:null,
"name":"count_2",
"extra_info":{},
"id":null,
"target":null,
"preview_display":true,
"display_name":"count_2",
base_type:"type/*"
source:
}
```
|
process
|
multiple aggregations of the same type are missing correct cols metadata similar to but when two aggregations of the same type are specified only the first has the correct cols metadata in particular base type special type and source are incorrect example query json source table aggregation breakout as day expected data cols and data cols json description null table id null special type type number name count source aggregation extra info id null target null display name count base type type integer description null table id null special type type number name count source aggregation extra info id null target null display name count base type type integer actual data cols and data cols json description null table id null special type type number name count source aggregation extra info id null target null display name count base type type integer description null table id null special type null name count extra info id null target null preview display true display name count base type type source
| 1
|
4,130
| 7,086,165,366
|
IssuesEvent
|
2018-01-11 13:43:46
|
rogerthat-platform/rogerthat-android-client
|
https://api.github.com/repos/rogerthat-platform/rogerthat-android-client
|
closed
|
Apps with a custom home screen
|
priority_critical process_duplicate type_feature
|
- [ ] Use homescreen_style "branding" in `build.yaml`
- [ ] After registration, show progress bar until main service, all js embeddings and home branding are available.
- [ ] Implement api to read news items from inside a branding
- Security: return all news items to the main service. Other services should only receive their news items.
- [ ] Triggering of badge number callbacks
|
1.0
|
Apps with a custom home screen - - [ ] Use homescreen_style "branding" in `build.yaml`
- [ ] After registration, show progress bar until main service, all js embeddings and home branding are available.
- [ ] Implement api to read news items from inside a branding
- Security: return all news items to the main service. Other services should only receive their news items.
- [ ] Triggering of badge number callbacks
|
process
|
apps with a custom home screen use homescreen style branding in build yaml after registration show progress bar until main service all js embeddings and home branding are available implement api to read news items from inside a branding security return all news items to the main service other services should only receive their news items triggering of badge number callbacks
| 1
|
17,602
| 23,425,325,115
|
IssuesEvent
|
2022-08-14 09:55:08
|
Battle-s/battle-school-backend
|
https://api.github.com/repos/Battle-s/battle-school-backend
|
opened
|
[FEAT] 시즌 생성 및 조회
|
feature :computer: processing :hourglass_flowing_sand:
|
## 설명
> 시즌 생성 및 조회
## 체크사항
- [ ] 시즌 엔티티 및 repo 생성
- [ ] 시즌 서비스 - crud
## 참고자료
## 관련 논의
|
1.0
|
[FEAT] 시즌 생성 및 조회 - ## 설명
> 시즌 생성 및 조회
## 체크사항
- [ ] 시즌 엔티티 및 repo 생성
- [ ] 시즌 서비스 - crud
## 참고자료
## 관련 논의
|
process
|
시즌 생성 및 조회 설명 시즌 생성 및 조회 체크사항 시즌 엔티티 및 repo 생성 시즌 서비스 crud 참고자료 관련 논의
| 1
|
78,465
| 27,533,147,127
|
IssuesEvent
|
2023-03-07 00:13:08
|
openzfs/zfs
|
https://api.github.com/repos/openzfs/zfs
|
closed
|
Performance degradation in NVMe L2ARC configured ZFS upgrading from 2.1.4 to 2.1.6
|
Type: Defect
|
<!-- Please fill out the following template, which will help other contributors address your issue. -->
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please check our issue tracker before opening a new issue.
Additional valuable information can be found in the OpenZFS documentation
and mailing list archives.
Please fill in as much of the template as possible.
-->
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Amazon Linux 2
Distribution Version | Amazon Linux 2 AMI 2.0.20230119.1 x86_64
Kernel Version | 5.10.165-143.735.amzn2.x86_64
Architecture | x86_64
OpenZFS Version | 2.1.6
<!--
Command to find OpenZFS version:
zfs version
Commands to find kernel version:
uname -r # Linux
freebsd-version -r # FreeBSD
-->
### Describe the problem you're observing
Upgrading from ZFS 2.1.4 to ZFS 2.1.6, I'm seeing increased read latency and reduced bandwidth, IOPS.
I bisected it to the following commit: [39e038cdc61a9af2549fa3ed1f42c16dbcbe4ca6](https://github.com/openzfs/zfs/pull/12285/commits/39e038cdc61a9af2549fa3ed1f42c16dbcbe4ca6)
I have the l2arc_exclude_special tunable set to 0 as is the default. Reverting this commit seems to fix the issue. The issue is only affecting performance for a NVMe L2ARC configured ZFS. ZFS without L2ARC is unaffected.
|
1.0
|
Performance degradation in NVMe L2ARC configured ZFS upgrading from 2.1.4 to 2.1.6 - <!-- Please fill out the following template, which will help other contributors address your issue. -->
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please check our issue tracker before opening a new issue.
Additional valuable information can be found in the OpenZFS documentation
and mailing list archives.
Please fill in as much of the template as possible.
-->
### System information
<!-- add version after "|" character -->
Type | Version/Name
--- | ---
Distribution Name | Amazon Linux 2
Distribution Version | Amazon Linux 2 AMI 2.0.20230119.1 x86_64
Kernel Version | 5.10.165-143.735.amzn2.x86_64
Architecture | x86_64
OpenZFS Version | 2.1.6
<!--
Command to find OpenZFS version:
zfs version
Commands to find kernel version:
uname -r # Linux
freebsd-version -r # FreeBSD
-->
### Describe the problem you're observing
Upgrading from ZFS 2.1.4 to ZFS 2.1.6, I'm seeing increased read latency and reduced bandwidth, IOPS.
I bisected it to the following commit: [39e038cdc61a9af2549fa3ed1f42c16dbcbe4ca6](https://github.com/openzfs/zfs/pull/12285/commits/39e038cdc61a9af2549fa3ed1f42c16dbcbe4ca6)
I have the l2arc_exclude_special tunable set to 0 as is the default. Reverting this commit seems to fix the issue. The issue is only affecting performance for a NVMe L2ARC configured ZFS. ZFS without L2ARC is unaffected.
|
non_process
|
performance degradation in nvme configured zfs upgrading from to thank you for reporting an issue important please check our issue tracker before opening a new issue additional valuable information can be found in the openzfs documentation and mailing list archives please fill in as much of the template as possible system information type version name distribution name amazon linux distribution version amazon linux ami kernel version architecture openzfs version command to find openzfs version zfs version commands to find kernel version uname r linux freebsd version r freebsd describe the problem you re observing upgrading from zfs to zfs i m seeing increased read latency and reduced bandwidth iops i bisected it to the following commit i have the exclude special tunable set to as is the default reverting this commit seems to fix the issue the issue is only affecting performance for a nvme configured zfs zfs without is unaffected
| 0
|
63,695
| 8,691,141,700
|
IssuesEvent
|
2018-12-04 00:01:29
|
CuBoulder/express
|
https://api.github.com/repos/CuBoulder/express
|
closed
|
As a SO/CE I would like to place interactive blocks on my website
|
Epic Still Open at 3.0 evaluate-1:Director Approval evaluate-1:More Information evaluate-2:Accessibility evaluate-2:Documentation Needs evaluate-2:Support Concerns evaluate-2:Usability
|
## Context
Good afternoon,
Do we have the feature in Drupal that allows us to hover a box and for text to appear like so:
http://alumni.umich.edu/
Thank you,
Janeyl

|
1.0
|
As a SO/CE I would like to place interactive blocks on my website - ## Context
Good afternoon,
Do we have the feature in Drupal that allows us to hover a box and for text to appear like so:
http://alumni.umich.edu/
Thank you,
Janeyl

|
non_process
|
as a so ce i would like to place interactive blocks on my website context good afternoon do we have the feature in drupal that allows us to hover a box and for text to appear like so thank you janeyl
| 0
|
21,660
| 30,107,792,392
|
IssuesEvent
|
2023-06-30 04:04:12
|
open-telemetry/opentelemetry-collector-contrib
|
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
|
closed
|
Rename websocketprocessor to RemoteObserverProcessor
|
enhancement needs triage processor/websocket
|
### Component(s)
processor/websocket
### Is your feature request related to a problem? Please describe.
As discussed in #19633, the websocketprocessor name is too broad and confusing, and doesn't characterize the intent of use to the satisfaction of the community.
### Describe the solution you'd like
Rename the websocketprocessor to RemoteObserverProcessor so its use becomes clear.
### Describe alternatives you've considered
_No response_
### Additional context
_No response_
|
1.0
|
Rename websocketprocessor to RemoteObserverProcessor - ### Component(s)
processor/websocket
### Is your feature request related to a problem? Please describe.
As discussed in #19633, the websocketprocessor name is too broad and confusing, and doesn't characterize the intent of use to the satisfaction of the community.
### Describe the solution you'd like
Rename the websocketprocessor to RemoteObserverProcessor so its use becomes clear.
### Describe alternatives you've considered
_No response_
### Additional context
_No response_
|
process
|
rename websocketprocessor to remoteobserverprocessor component s processor websocket is your feature request related to a problem please describe as discussed in the websocketprocessor name is too broad and confusing and doesn t characterize the intent of use to the satisfaction of the community describe the solution you d like rename the websocketprocessor to remoteobserverprocessor so its use becomes clear describe alternatives you ve considered no response additional context no response
| 1
|
6,356
| 9,415,072,604
|
IssuesEvent
|
2019-04-10 11:46:00
|
AmpersandTarski/Ampersand
|
https://api.github.com/repos/AmpersandTarski/Ampersand
|
opened
|
stack install fails on MacBook
|
OSX priority:normal software process
|
#### Version of ampersand that was used
This problem occurred on commit c0239e76195cbf7692b5bb7e15dbcdb7aafd60cf, on the development branch of the github Ampersand repository.
#### What I expected
Since the purpose of stack is to build in platform independent ways, I expected "stack install" to build Ampersand for me.
#### What happened instead
The command "stack install" fails quite quickly:
<img width="682" alt="Screenshot 2019-04-10 at 13 30 52" src="https://user-images.githubusercontent.com/275768/55875911-8bc9f000-5b96-11e9-9995-c06879796b43.png">
Apparently, it uses an executable meant for Windows.
#### Steps to reproduce
1.
2.
3.
4.
#### Screenshot / Video
#### Context / Source of ampersand script
<!-- Optional: share your script if possible. It helps us reproduce the problem. Please try to keep the scripts tiny
We'd also love to know how you found the bug: #dogfooding, #manual-testing, #automated-testing, or #user-report if applicable.
If requesting a new feature, explain why you'd like to see it added.
-->
|
1.0
|
stack install fails on MacBook - #### Version of ampersand that was used
This problem occurred on commit c0239e76195cbf7692b5bb7e15dbcdb7aafd60cf, on the development branch of the github Ampersand repository.
#### What I expected
Since the purpose of stack is to build in platform independent ways, I expected "stack install" to build Ampersand for me.
#### What happened instead
The command "stack install" fails quite quickly:
<img width="682" alt="Screenshot 2019-04-10 at 13 30 52" src="https://user-images.githubusercontent.com/275768/55875911-8bc9f000-5b96-11e9-9995-c06879796b43.png">
Apparently, it uses an executable meant for Windows.
#### Steps to reproduce
1.
2.
3.
4.
#### Screenshot / Video
#### Context / Source of ampersand script
<!-- Optional: share your script if possible. It helps us reproduce the problem. Please try to keep the scripts tiny
We'd also love to know how you found the bug: #dogfooding, #manual-testing, #automated-testing, or #user-report if applicable.
If requesting a new feature, explain why you'd like to see it added.
-->
|
process
|
stack install fails on macbook version of ampersand that was used this problem occurred on commit on the development branch of the github ampersand repository what i expected since the purpose of stack is to build in platform independent ways i expected stack install to build ampersand for me what happened instead the command stack install fails quite quickly img width alt screenshot at src apparently it uses an executable meant for windows steps to reproduce screenshot video context source of ampersand script optional share your script if possible it helps us reproduce the problem please try to keep the scripts tiny we d also love to know how you found the bug dogfooding manual testing automated testing or user report if applicable if requesting a new feature explain why you d like to see it added
| 1
|
21,727
| 30,235,103,069
|
IssuesEvent
|
2023-07-06 09:40:13
|
mrdoob/three.js
|
https://api.github.com/repos/mrdoob/three.js
|
closed
|
MaskPass breaks background update for transmissive objects
|
Bug Post-processing
|
Currently there are the following [lines of code in the MaskPass class](https://github.com/mrdoob/three.js/blob/42c59b4d516fd204f5f760e1f7daefbbbf18c49c/examples/jsm/postprocessing/MaskPass.js#L26)
```js
state.buffers.color.setMask( false );
state.buffers.depth.setMask( false );
```
Both values are set to `false`, but these operations aren't reverted at the end of the pass.
As the result, if after the scene is rendered you remove the scene background, it's not removed for transmissive objects.
What's more, if there are non-transmissive parts in a transmissive model and it's being rotated, then the non-transmissive parts leave a trail on the background.
1. A box mask is applied to a scene with some background and a transmissive object. Note that there is a small non-transmissive box in the model.

2. The scene background is removed, but it's still there for the transmissive object.

3. When the object is rotated, the small opaque black box leaves a trail on the background.

**Some code**
```js
function createBox(scale = 1, x = 0, y = 0, z = 0) {
const geometry = new BoxGeometry(scale , scale , scale);
const mesh = new Mesh(geometry);
mesh.material = new MeshBasicMaterial({ color: 0x000000 });
mesh.position.x = x;
mesh.position.y = y;
mesh.position.z = z;
return mesh;
}
const maskScene = new Scene();
maskScene.add(createBox(3));
// load a model and add an opaque box to it
gltf.scene.add(createBox(0.1, -0.5, 0.5, 0));
scene.add(gltf.scene);
const maskRenderTarget = new WebGLRenderTarget(500, 500, {
stencilBuffer: true,
samples: 8,
});
const composer = new EffectComposer(renderer, maskRenderTarget);
composer.setPixelRatio(window.devicePixelRatio);
// passes
composer.addPass(new ClearPass());
composer.addPass(new MaskPass(maskScene, camera));
const renderPass = new RenderPass(scene, camera);
renderPass.clear = false;
composer.addPass(renderPass);
composer.addPass(new ClearMaskPass());
composer.addPass(new ShaderPass(GammaCorrectionShader));
```
**Probable solution**
If the following code
```js
state.buffers.color.setMask( true );
state.buffers.depth.setMask( true );
```
is executed before `RenderPass` (e.g. at the end of `MaskPass`), then the problem disappears.
For this very issue only the first line is enough.
|
1.0
|
MaskPass breaks background update for transmissive objects - Currently there are the following [lines of code in the MaskPass class](https://github.com/mrdoob/three.js/blob/42c59b4d516fd204f5f760e1f7daefbbbf18c49c/examples/jsm/postprocessing/MaskPass.js#L26)
```js
state.buffers.color.setMask( false );
state.buffers.depth.setMask( false );
```
Both values are set to `false`, but these operations aren't reverted at the end of the pass.
As the result, if after the scene is rendered you remove the scene background, it's not removed for transmissive objects.
What's more, if there are non-transmissive parts in a transmissive model and it's being rotated, then the non-transmissive parts leave a trail on the background.
1. A box mask is applied to a scene with some background and a transmissive object. Note that there is a small non-transmissive box in the model.

2. The scene background is removed, but it's still there for the transmissive object.

3. When the object is rotated, the small opaque black box leaves a trail on the background.

**Some code**
```js
function createBox(scale = 1, x = 0, y = 0, z = 0) {
const geometry = new BoxGeometry(scale , scale , scale);
const mesh = new Mesh(geometry);
mesh.material = new MeshBasicMaterial({ color: 0x000000 });
mesh.position.x = x;
mesh.position.y = y;
mesh.position.z = z;
return mesh;
}
const maskScene = new Scene();
maskScene.add(createBox(3));
// load a model and add an opaque box to it
gltf.scene.add(createBox(0.1, -0.5, 0.5, 0));
scene.add(gltf.scene);
const maskRenderTarget = new WebGLRenderTarget(500, 500, {
stencilBuffer: true,
samples: 8,
});
const composer = new EffectComposer(renderer, maskRenderTarget);
composer.setPixelRatio(window.devicePixelRatio);
// passes
composer.addPass(new ClearPass());
composer.addPass(new MaskPass(maskScene, camera));
const renderPass = new RenderPass(scene, camera);
renderPass.clear = false;
composer.addPass(renderPass);
composer.addPass(new ClearMaskPass());
composer.addPass(new ShaderPass(GammaCorrectionShader));
```
**Probable solution**
If the following code
```js
state.buffers.color.setMask( true );
state.buffers.depth.setMask( true );
```
is executed before `RenderPass` (e.g. at the end of `MaskPass`), then the problem disappears.
For this very issue only the first line is enough.
|
process
|
maskpass breaks background update for transmissive objects currently there are the following js state buffers color setmask false state buffers depth setmask false both values are set to false but these operations aren t reverted at the end of the pass as the result if after the scene is rendered you remove the scene background it s not removed for transmissive objects what s more if there are non transmissive parts in a transmissive model and it s being rotated then the non transmissive parts leave a trail on the background a box mask is applied to a scene with some background and a transmissive object note that there is a small non transmissive box in the model the scene background is removed but it s still there for the transmissive object when the object is rotated the small opaque black box leaves a trail on the background some code js function createbox scale x y z const geometry new boxgeometry scale scale scale const mesh new mesh geometry mesh material new meshbasicmaterial color mesh position x x mesh position y y mesh position z z return mesh const maskscene new scene maskscene add createbox load a model and add an opaque box to it gltf scene add createbox scene add gltf scene const maskrendertarget new webglrendertarget stencilbuffer true samples const composer new effectcomposer renderer maskrendertarget composer setpixelratio window devicepixelratio passes composer addpass new clearpass composer addpass new maskpass maskscene camera const renderpass new renderpass scene camera renderpass clear false composer addpass renderpass composer addpass new clearmaskpass composer addpass new shaderpass gammacorrectionshader probable solution if the following code js state buffers color setmask true state buffers depth setmask true is executed before renderpass e g at the end of maskpass then the problem disappears for this very issue only the first line is enough
| 1
|
117,851
| 25,203,539,006
|
IssuesEvent
|
2022-11-13 12:05:44
|
leo-editor/leo-editor
|
https://api.github.com/repos/leo-editor/leo-editor
|
closed
|
Fix problems with ftm.set_widgets and sftm.init_widgets
|
Code
|
From an email from Félix. There are several problems with FindtabManager.set_widgets_from_dict and StringFindTabManager.init_widgets:
- sftm.init_widgets does not handle the "File Only" option.
- Neither method defines "val" after the comment `# Ensure one radio button is set.`
- Both methods should fall back to "Entire Outline" instead of "File Only"!
|
1.0
|
Fix problems with ftm.set_widgets and sftm.init_widgets - From an email from Félix. There are several problems with FindtabManager.set_widgets_from_dict and StringFindTabManager.init_widgets:
- sftm.init_widgets does not handle the "File Only" option.
- Neither method defines "val" after the comment `# Ensure one radio button is set.`
- Both methods should fall back to "Entire Outline" instead of "File Only"!
|
non_process
|
fix problems with ftm set widgets and sftm init widgets from an email from félix there are several problems with findtabmanager set widgets from dict and stringfindtabmanager init widgets sftm init widgets does not handle the file only option neither method defines val after the comment ensure one radio button is set both methods should fall back to entire outline instead of file only
| 0
|
323,582
| 27,738,234,017
|
IssuesEvent
|
2023-03-15 12:44:37
|
Jersey-Electricity-plc/intsys-issues
|
https://api.github.com/repos/Jersey-Electricity-plc/intsys-issues
|
closed
|
User Setup for NTS_D1\SLEG message when trying to create a quote
|
bug readyToTest
|
Trying to create a quote for HD0018781 but the bellow message appears. I have looked at user set up but my user ID is different to everyone else.
Mine is SLEG not NTS_D1\SLEG, apologies if I'm missing something.

|
1.0
|
User Setup for NTS_D1\SLEG message when trying to create a quote - Trying to create a quote for HD0018781 but the bellow message appears. I have looked at user set up but my user ID is different to everyone else.
Mine is SLEG not NTS_D1\SLEG, apologies if I'm missing something.

|
non_process
|
user setup for nts sleg message when trying to create a quote trying to create a quote for but the bellow message appears i have looked at user set up but my user id is different to everyone else mine is sleg not nts sleg apologies if i m missing something
| 0
|
12,038
| 14,738,709,423
|
IssuesEvent
|
2021-01-07 05:31:09
|
kdjstudios/SABillingGitlab
|
https://api.github.com/repos/kdjstudios/SABillingGitlab
|
closed
|
Spill Center Language Line
|
anc-ops anc-process anp-1.5 ant-support
|
In GitLab by @kdjstudios on Jul 11, 2018, 14:00
**Submitted by:** "Kimberly Gagner" <kimberly.gagner@answernet.com>
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-07-11-14887/conversation
**Server:** Internal
**Client/Site:** 123
**Account:** Multiple
**Issue:**
I am reaching out to you to ask for your help in fixing the $50 fees that are appearing on some of the Spill Center language accounts. I know why this has happened and I will explain.
When the language accounts first were set up in SAB they were set up as individual accounts. Now, the language accounts are all linked to E0840 and billed from that account. Since the link was added and the $50.00 charges were already showing on the invoice before the accounts were linked is why I cannot remove the charges. I have tried entering staged fees individually in each one and it is not working correctly. I need someone with the ability to go into the following accounts and remove those $50.00 charges. Here is the list:
E0844 – Spill Center Cantonese
E0842 – Spill Center Czech
E0852 – Spill Center Flemish
E0846 – Spill Center Japanese
E0851 – Spill Center Korean
E0853 – Spill Center Polish
E0854 – Spill Center – Portuguese
E0849 – Spill Center – Turkish
Another reason this needs to be corrected is it is appearing on my aging report.
Kyle, I attached a screenshot as well. If you have questions, please call me at x12301.
|
1.0
|
Spill Center Language Line - In GitLab by @kdjstudios on Jul 11, 2018, 14:00
**Submitted by:** "Kimberly Gagner" <kimberly.gagner@answernet.com>
**Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-07-11-14887/conversation
**Server:** Internal
**Client/Site:** 123
**Account:** Multiple
**Issue:**
I am reaching out to you to ask for your help in fixing the $50 fees that are appearing on some of the Spill Center language accounts. I know why this has happened and I will explain.
When the language accounts first were set up in SAB they were set up as individual accounts. Now, the language accounts are all linked to E0840 and billed from that account. Since the link was added and the $50.00 charges were already showing on the invoice before the accounts were linked is why I cannot remove the charges. I have tried entering staged fees individually in each one and it is not working correctly. I need someone with the ability to go into the following accounts and remove those $50.00 charges. Here is the list:
E0844 – Spill Center Cantonese
E0842 – Spill Center Czech
E0852 – Spill Center Flemish
E0846 – Spill Center Japanese
E0851 – Spill Center Korean
E0853 – Spill Center Polish
E0854 – Spill Center – Portuguese
E0849 – Spill Center – Turkish
Another reason this needs to be corrected is it is appearing on my aging report.
Kyle, I attached a screenshot as well. If you have questions, please call me at x12301.
|
process
|
spill center language line in gitlab by kdjstudios on jul submitted by kimberly gagner helpdesk server internal client site account multiple issue i am reaching out to you to ask for your help in fixing the fees that are appearing on some of the spill center language accounts i know why this has happened and i will explain when the language accounts first were set up in sab they were set up as individual accounts now the language accounts are all linked to and billed from that account since the link was added and the charges were already showing on the invoice before the accounts were linked is why i cannot remove the charges i have tried entering staged fees individually in each one and it is not working correctly i need someone with the ability to go into the following accounts and remove those charges here is the list – spill center cantonese – spill center czech – spill center flemish – spill center japanese – spill center korean – spill center polish – spill center – portuguese – spill center – turkish another reason this needs to be corrected is it is appearing on my aging report kyle i attached a screenshot as well if you have questions please call me at
| 1
|
8,438
| 11,599,117,911
|
IssuesEvent
|
2020-02-25 01:10:38
|
MicrosoftDocs/azure-docs
|
https://api.github.com/repos/MicrosoftDocs/azure-docs
|
closed
|
-ThrottleLimit
|
Pri2 automation/svc cxp doc-enhancement process-automation/subsvc triaged
|
It's not clear what -ThrottleLimit object value means.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 805a6236-70b7-7dd5-ac86-eea6efceff3a
* Version Independent ID: e8078e34-bdf0-32b1-fac4-550091f2a06a
* Content: [Learning PowerShell Workflow for Azure Automation](https://docs.microsoft.com/en-us/azure/automation/automation-powershell-workflow#feedback)
* Content Source: [articles/automation/automation-powershell-workflow.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-powershell-workflow.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @MGoedtel
* Microsoft Alias: **magoedte**
|
1.0
|
-ThrottleLimit - It's not clear what -ThrottleLimit object value means.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 805a6236-70b7-7dd5-ac86-eea6efceff3a
* Version Independent ID: e8078e34-bdf0-32b1-fac4-550091f2a06a
* Content: [Learning PowerShell Workflow for Azure Automation](https://docs.microsoft.com/en-us/azure/automation/automation-powershell-workflow#feedback)
* Content Source: [articles/automation/automation-powershell-workflow.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-powershell-workflow.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @MGoedtel
* Microsoft Alias: **magoedte**
|
process
|
throttlelimit it s not clear what throttlelimit object value means document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login mgoedtel microsoft alias magoedte
| 1
|
6,636
| 9,745,796,601
|
IssuesEvent
|
2019-06-03 10:30:34
|
EthVM/EthVM
|
https://api.github.com/repos/EthVM/EthVM
|
opened
|
Parametrize window period on Kafka Processing
|
bug enhancement priority:high project:processing
|
* **I'm submitting a ...**
- [X] feature request
* **Feature Request**
- In processing MainNet we obtained the following exception:
```
[2019-06-03 07:02:48,998] WARN WorkerSinkTask{id=postgres-block-sink-0} Commit of offsets timed out (org.apache.kafka.connect.runtime.WorkerSinkTask)
```
Which means that if for some reason we wait too long to obtain new data, the windows are expired. We need to parametrize that in our Kafka processing with an Env variable (so different networks can have different settings).
I'll mark this as an enhancement and as a bug, as we need this change for processing properly MainNet.
|
1.0
|
Parametrize window period on Kafka Processing - * **I'm submitting a ...**
- [X] feature request
* **Feature Request**
- In processing MainNet we obtained the following exception:
```
[2019-06-03 07:02:48,998] WARN WorkerSinkTask{id=postgres-block-sink-0} Commit of offsets timed out (org.apache.kafka.connect.runtime.WorkerSinkTask)
```
Which means that if for some reason we wait too long to obtain new data, the windows are expired. We need to parametrize that in our Kafka processing with an Env variable (so different networks can have different settings).
I'll mark this as an enhancement and as a bug, as we need this change for processing properly MainNet.
|
process
|
parametrize window period on kafka processing i m submitting a feature request feature request in processing mainnet we obtained the following exception warn workersinktask id postgres block sink commit of offsets timed out org apache kafka connect runtime workersinktask which means that if for some reason we wait too long to obtain new data the windows are expired we need to parametrize that in our kafka processing with an env variable so different networks can have different settings i ll mark this as an enhancement and as a bug as we need this change for processing properly mainnet
| 1
|
10,291
| 2,939,353,831
|
IssuesEvent
|
2015-07-01 15:52:04
|
WikiWatershed/model-my-watershed
|
https://api.github.com/repos/WikiWatershed/model-my-watershed
|
opened
|
iPad Issues
|
design
|
Taken on iPad Mini 2 running iOS 8.3.
* The modifications dropdowns has width issues:

* In some cases, the modification thumbnails are offscreen and the user cannot scroll below to see them:

* Unlike behavior on Android, on an iPad tapping once simulates a Hover, and tapping twice simulates a Click (selection). Thus the user must tap twice on any modification to select it, which is not very obvious. Also, when the user is in the "Draw" mode, there is no visual feedback for this (unlike on the desktop where the cursor changes and has an attached tooltip).
|
1.0
|
iPad Issues - Taken on iPad Mini 2 running iOS 8.3.
* The modifications dropdowns has width issues:

* In some cases, the modification thumbnails are offscreen and the user cannot scroll below to see them:

* Unlike behavior on Android, on an iPad tapping once simulates a Hover, and tapping twice simulates a Click (selection). Thus the user must tap twice on any modification to select it, which is not very obvious. Also, when the user is in the "Draw" mode, there is no visual feedback for this (unlike on the desktop where the cursor changes and has an attached tooltip).
|
non_process
|
ipad issues taken on ipad mini running ios the modifications dropdowns has width issues in some cases the modification thumbnails are offscreen and the user cannot scroll below to see them unlike behavior on android on an ipad tapping once simulates a hover and tapping twice simulates a click selection thus the user must tap twice on any modification to select it which is not very obvious also when the user is in the draw mode there is no visual feedback for this unlike on the desktop where the cursor changes and has an attached tooltip
| 0
|
166,375
| 26,345,623,824
|
IssuesEvent
|
2023-01-10 21:46:41
|
gotogether-s/gotogether-s
|
https://api.github.com/repos/gotogether-s/gotogether-s
|
closed
|
SignUp
|
design
|
## About
Update style of signup page
## To do list
- [x] Sort import lines of signup page
- [x] Organize code lines of signup page
- [x] Reduce paddingBottom from 1rem to 0.5rem in signin page
- [x] Reduce paddingTop value of error message in signin page
- [x] Reduce marginBottom value of input wrapper box
- [x] Remove marginTop value of signup link
- [x] Reduce default marginBottom value to 2rem
- [x] Add style to Button on hover
- [x] Replace HTML with MUI, Update style and Delete SignUp.scss
- [x] Reduce size of icon and Increase fontSize of page title
|
1.0
|
SignUp - ## About
Update style of signup page
## To do list
- [x] Sort import lines of signup page
- [x] Organize code lines of signup page
- [x] Reduce paddingBottom from 1rem to 0.5rem in signin page
- [x] Reduce paddingTop value of error message in signin page
- [x] Reduce marginBottom value of input wrapper box
- [x] Remove marginTop value of signup link
- [x] Reduce default marginBottom value to 2rem
- [x] Add style to Button on hover
- [x] Replace HTML with MUI, Update style and Delete SignUp.scss
- [x] Reduce size of icon and Increase fontSize of page title
|
non_process
|
signup about update style of signup page to do list sort import lines of signup page organize code lines of signup page reduce paddingbottom from to in signin page reduce paddingtop value of error message in signin page reduce marginbottom value of input wrapper box remove margintop value of signup link reduce default marginbottom value to add style to button on hover replace html with mui update style and delete signup scss reduce size of icon and increase fontsize of page title
| 0
|
818,783
| 30,704,318,196
|
IssuesEvent
|
2023-07-27 04:09:26
|
ballerina-platform/ballerina-lang
|
https://api.github.com/repos/ballerina-platform/ballerina-lang
|
closed
|
[Bug]: NPE in ComputedResourceAccessSegmentNodeContext
|
Type/Bug Priority/High Team/LanguageServer Points/1 Area/Completion userCategory/Editor
|
### Description
In the following scenario, when we type something after `"path" a`, a null pointer exception is thrown.
https://github.com/ballerina-platform/ballerina-lang/assets/61020198/3b61c5ad-b901-46d2-83f5-d6370c7af104
### Steps to Reproduce
```
client class ClientClass {
resource function accessor ["path" a] () {
}
}
public function main() {
var cl = new ClientClass();
cl ->/["path"].accessor;
}
```
### Affected Version(s)
_No response_
### OS, DB, other environment details and versions
_No response_
### Related area
-> Editor
### Related issue(s) (optional)
_No response_
### Suggested label(s) (optional)
_No response_
### Suggested assignee(s) (optional)
_No response_
|
1.0
|
[Bug]: NPE in ComputedResourceAccessSegmentNodeContext - ### Description
In the following scenario, when we type something after `"path" a`, a null pointer exception is thrown.
https://github.com/ballerina-platform/ballerina-lang/assets/61020198/3b61c5ad-b901-46d2-83f5-d6370c7af104
### Steps to Reproduce
```
client class ClientClass {
resource function accessor ["path" a] () {
}
}
public function main() {
var cl = new ClientClass();
cl ->/["path"].accessor;
}
```
### Affected Version(s)
_No response_
### OS, DB, other environment details and versions
_No response_
### Related area
-> Editor
### Related issue(s) (optional)
_No response_
### Suggested label(s) (optional)
_No response_
### Suggested assignee(s) (optional)
_No response_
|
non_process
|
npe in computedresourceaccesssegmentnodecontext description in the following scenario when we type something after path a a null pointer exception is thrown steps to reproduce client class clientclass resource function accessor public function main var cl new clientclass cl accessor affected version s no response os db other environment details and versions no response related area editor related issue s optional no response suggested label s optional no response suggested assignee s optional no response
| 0
|
2,915
| 5,909,019,253
|
IssuesEvent
|
2017-05-19 22:12:06
|
metabase/metabase
|
https://api.github.com/repos/metabase/metabase
|
opened
|
Second date breakout doesn't have `unit` in the result `col`
|
Query Processor
|
If I breakout by 2 date columns only the first has a `unit` in the response `cols`.
This causes headers in pivot tables to not be formatted correctly, and probably various other issues.
```json
{
"data": {
"cols" : [
{
"source" : "breakout",
"schema_name" : "PUBLIC",
"extra_info" : {},
"special_type" : null,
"name" : "CREATED_AT",
"fk_field_id" : null,
"visibility_type" : "normal",
"display_name" : "Created At",
"target" : null,
"base_type" : "type/DateTime",
"id" : 1,
"description" : "The date and time an order was submitted.",
"table_id" : 1,
"unit" : "month"
},
{
"preview_display" : true,
"display_name" : "CREATED_AT_2",
"target" : null,
"id" : null,
"base_type" : "type/DateTime",
"table_id" : null,
"description" : null,
"source" : "breakout",
"extra_info" : {},
"special_type" : null,
"name" : "CREATED_AT_2"
},
{
"extra_info" : {},
"special_type" : "type/Number",
"name" : "count",
"source" : "aggregation",
"base_type" : "type/Integer",
"id" : null,
"description" : null,
"table_id" : null,
"target" : null,
"display_name" : "count"
}
]
},
"json_query" : {
"query" : {
"source_table" : 1,
"aggregation" : [
[
"count"
]
],
"breakout" : [
[
"datetime-field",
[
"field-id",
1
],
"as",
"month"
],
[
"datetime-field",
[
"fk->",
3,
22
],
"as",
"month"
]
]
},
"type" : "query",
"parameters" : [],
"database" : 1,
"constraints" : {
"max-results-bare-rows" : 2000,
"max-results" : 10000
}
}
}
```
|
1.0
|
Second date breakout doesn't have `unit` in the result `col` - If I breakout by 2 date columns only the first has a `unit` in the response `cols`.
This causes headers in pivot tables to not be formatted correctly, and probably various other issues.
```json
{
"data": {
"cols" : [
{
"source" : "breakout",
"schema_name" : "PUBLIC",
"extra_info" : {},
"special_type" : null,
"name" : "CREATED_AT",
"fk_field_id" : null,
"visibility_type" : "normal",
"display_name" : "Created At",
"target" : null,
"base_type" : "type/DateTime",
"id" : 1,
"description" : "The date and time an order was submitted.",
"table_id" : 1,
"unit" : "month"
},
{
"preview_display" : true,
"display_name" : "CREATED_AT_2",
"target" : null,
"id" : null,
"base_type" : "type/DateTime",
"table_id" : null,
"description" : null,
"source" : "breakout",
"extra_info" : {},
"special_type" : null,
"name" : "CREATED_AT_2"
},
{
"extra_info" : {},
"special_type" : "type/Number",
"name" : "count",
"source" : "aggregation",
"base_type" : "type/Integer",
"id" : null,
"description" : null,
"table_id" : null,
"target" : null,
"display_name" : "count"
}
]
},
"json_query" : {
"query" : {
"source_table" : 1,
"aggregation" : [
[
"count"
]
],
"breakout" : [
[
"datetime-field",
[
"field-id",
1
],
"as",
"month"
],
[
"datetime-field",
[
"fk->",
3,
22
],
"as",
"month"
]
]
},
"type" : "query",
"parameters" : [],
"database" : 1,
"constraints" : {
"max-results-bare-rows" : 2000,
"max-results" : 10000
}
}
}
```
|
process
|
second date breakout doesn t have unit in the result col if i breakout by date columns only the first has a unit in the response cols this causes headers in pivot tables to not be formatted correctly and probably various other issues json data cols source breakout schema name public extra info special type null name created at fk field id null visibility type normal display name created at target null base type type datetime id description the date and time an order was submitted table id unit month preview display true display name created at target null id null base type type datetime table id null description null source breakout extra info special type null name created at extra info special type type number name count source aggregation base type type integer id null description null table id null target null display name count json query query source table aggregation count breakout datetime field field id as month datetime field fk as month type query parameters database constraints max results bare rows max results
| 1
|
2,284
| 5,108,914,081
|
IssuesEvent
|
2017-01-05 19:10:51
|
jlm2017/jlm-video-subtitles
|
https://api.github.com/repos/jlm2017/jlm-video-subtitles
|
opened
|
[Subtitles] [FR] VOEUX DE JEAN-LUC MÉLENCHON - UNE FRANCE AU SERVICE DE LA PAIX
|
Language: French Process: Someone is working on this issue Process: [1] Writing in progress
|
# Video title
VOEUX DE JEAN-LUC MÉLENCHON - UNE FRANCE AU SERVICE DE LA PAIX
# URL
https://www.youtube.com/watch?v=u8Lta7F4SgY
# Youtube subtitles language
Français
# Duration
33:56
# Subtitles URL
https://www.youtube.com/timedtext_editor?ref=player&tab=captions&lang=fr&v=u8Lta7F4SgY&action_mde_edit_form=1&ui=hd&bl=vmp
|
2.0
|
[Subtitles] [FR] VOEUX DE JEAN-LUC MÉLENCHON - UNE FRANCE AU SERVICE DE LA PAIX - # Video title
VOEUX DE JEAN-LUC MÉLENCHON - UNE FRANCE AU SERVICE DE LA PAIX
# URL
https://www.youtube.com/watch?v=u8Lta7F4SgY
# Youtube subtitles language
Français
# Duration
33:56
# Subtitles URL
https://www.youtube.com/timedtext_editor?ref=player&tab=captions&lang=fr&v=u8Lta7F4SgY&action_mde_edit_form=1&ui=hd&bl=vmp
|
process
|
voeux de jean luc mélenchon une france au service de la paix video title voeux de jean luc mélenchon une france au service de la paix url youtube subtitles language français duration subtitles url
| 1
|
19,888
| 26,334,712,506
|
IssuesEvent
|
2023-01-10 13:32:08
|
bazelbuild/bazel
|
https://api.github.com/repos/bazelbuild/bazel
|
closed
|
How can I make some build args as required and fail the bazel build when the arg is not provided?
|
P3 type: support / not a bug (process) team-OSS
|
How can I make some build args as required and fail the bazel build when the arg is not provided? I have a bazel build <target> --config=X, I need a way to fail the bazel build if X is not provided.
|
1.0
|
How can I make some build args as required and fail the bazel build when the arg is not provided? - How can I make some build args as required and fail the bazel build when the arg is not provided? I have a bazel build <target> --config=X, I need a way to fail the bazel build if X is not provided.
|
process
|
how can i make some build args as required and fail the bazel build when the arg is not provided how can i make some build args as required and fail the bazel build when the arg is not provided i have a bazel build config x i need a way to fail the bazel build if x is not provided
| 1
|
199,337
| 22,693,305,426
|
IssuesEvent
|
2022-07-05 01:10:38
|
TechnoConserve/personal_website
|
https://api.github.com/repos/TechnoConserve/personal_website
|
opened
|
jquery-3.0.0.min.js: 3 vulnerabilities (highest severity is: 6.1)
|
security vulnerability
|
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-3.0.0.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0/jquery.min.js</a></p>
<p>Path to dependency file: /photo_blog/templates/base.html</p>
<p>Path to vulnerable library: /photo_blog/templates/base.html</p>
<p>
</details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2020-11023](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-3.0.0.min.js | Direct | jquery - 3.5.0;jquery-rails - 4.4.0 | ❌ |
| [CVE-2020-11022](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-3.0.0.min.js | Direct | jQuery - 3.5.0 | ❌ |
| [CVE-2019-11358](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11358) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-3.0.0.min.js | Direct | 3.4.0 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-11023</summary>
### Vulnerable Library - <b>jquery-3.0.0.min.js</b></p>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0/jquery.min.js</a></p>
<p>Path to dependency file: /photo_blog/templates/base.html</p>
<p>Path to vulnerable library: /photo_blog/templates/base.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-3.0.0.min.js** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023>CVE-2020-11023</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440">https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jquery - 3.5.0;jquery-rails - 4.4.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-11022</summary>
### Vulnerable Library - <b>jquery-3.0.0.min.js</b></p>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0/jquery.min.js</a></p>
<p>Path to dependency file: /photo_blog/templates/base.html</p>
<p>Path to vulnerable library: /photo_blog/templates/base.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-3.0.0.min.js** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jQuery - 3.5.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2019-11358</summary>
### Vulnerable Library - <b>jquery-3.0.0.min.js</b></p>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0/jquery.min.js</a></p>
<p>Path to dependency file: /photo_blog/templates/base.html</p>
<p>Path to vulnerable library: /photo_blog/templates/base.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-3.0.0.min.js** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype.
<p>Publish Date: 2019-04-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11358>CVE-2019-11358</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358</a></p>
<p>Release Date: 2019-04-20</p>
<p>Fix Resolution: 3.4.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
|
True
|
jquery-3.0.0.min.js: 3 vulnerabilities (highest severity is: 6.1) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-3.0.0.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0/jquery.min.js</a></p>
<p>Path to dependency file: /photo_blog/templates/base.html</p>
<p>Path to vulnerable library: /photo_blog/templates/base.html</p>
<p>
</details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2020-11023](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-3.0.0.min.js | Direct | jquery - 3.5.0;jquery-rails - 4.4.0 | ❌ |
| [CVE-2020-11022](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-3.0.0.min.js | Direct | jQuery - 3.5.0 | ❌ |
| [CVE-2019-11358](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11358) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-3.0.0.min.js | Direct | 3.4.0 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-11023</summary>
### Vulnerable Library - <b>jquery-3.0.0.min.js</b></p>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0/jquery.min.js</a></p>
<p>Path to dependency file: /photo_blog/templates/base.html</p>
<p>Path to vulnerable library: /photo_blog/templates/base.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-3.0.0.min.js** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023>CVE-2020-11023</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440">https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jquery - 3.5.0;jquery-rails - 4.4.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-11022</summary>
### Vulnerable Library - <b>jquery-3.0.0.min.js</b></p>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0/jquery.min.js</a></p>
<p>Path to dependency file: /photo_blog/templates/base.html</p>
<p>Path to vulnerable library: /photo_blog/templates/base.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-3.0.0.min.js** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jQuery - 3.5.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2019-11358</summary>
### Vulnerable Library - <b>jquery-3.0.0.min.js</b></p>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0/jquery.min.js</a></p>
<p>Path to dependency file: /photo_blog/templates/base.html</p>
<p>Path to vulnerable library: /photo_blog/templates/base.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-3.0.0.min.js** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype.
<p>Publish Date: 2019-04-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11358>CVE-2019-11358</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358</a></p>
<p>Release Date: 2019-04-20</p>
<p>Fix Resolution: 3.4.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
|
non_process
|
jquery min js vulnerabilities highest severity is vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file photo blog templates base html path to vulnerable library photo blog templates base html vulnerabilities cve severity cvss dependency type fixed in remediation available medium jquery min js direct jquery jquery rails medium jquery min js direct jquery medium jquery min js direct details cve vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file photo blog templates base html path to vulnerable library photo blog templates base html dependency hierarchy x jquery min js vulnerable library found in base branch main vulnerability details in jquery versions greater than or equal to and before passing html containing elements from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery jquery rails step up your open source security game with mend cve vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file photo blog templates base html path to vulnerable library photo blog templates base html dependency hierarchy x jquery min js vulnerable library found in base branch main vulnerability details in jquery versions greater than or equal to and before passing html from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with mend cve vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file photo blog templates base html path to vulnerable library photo blog templates base html dependency hierarchy x jquery min js vulnerable library found in base branch main vulnerability details jquery before as used in drupal backdrop cms and other products mishandles jquery extend true because of object prototype pollution if an unsanitized source object contained an enumerable proto property it could extend the native object prototype publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
55,683
| 11,458,267,115
|
IssuesEvent
|
2020-02-07 02:43:28
|
joomla/joomla-cms
|
https://api.github.com/repos/joomla/joomla-cms
|
closed
|
Error in installing Hindi language extension.
|
No Code Attached Yet
|
### Steps to reproduce the issue
Go to system settings and try to install hindi language extenstion.
### Expected result
### Actual result

### System information (as much as possible)
version: 4.0.0-beta1-dev
### Additional comments
|
1.0
|
Error in installing Hindi language extension. - ### Steps to reproduce the issue
Go to system settings and try to install hindi language extenstion.
### Expected result
### Actual result

### System information (as much as possible)
version: 4.0.0-beta1-dev
### Additional comments
|
non_process
|
error in installing hindi language extension steps to reproduce the issue go to system settings and try to install hindi language extenstion expected result actual result system information as much as possible version dev additional comments
| 0
|
39,735
| 20,177,619,101
|
IssuesEvent
|
2022-02-10 15:41:44
|
JuliaRobotics/IncrementalInference.jl
|
https://api.github.com/repos/JuliaRobotics/IncrementalInference.jl
|
closed
|
Must add cache to CalcFactor
|
performance types factors
|
Must add cache to CalcFactor
_Originally posted by @dehann in https://github.com/JuliaRobotics/IncrementalInference.jl/issues/1462#issuecomment-1008447079_
|
True
|
Must add cache to CalcFactor - Must add cache to CalcFactor
_Originally posted by @dehann in https://github.com/JuliaRobotics/IncrementalInference.jl/issues/1462#issuecomment-1008447079_
|
non_process
|
must add cache to calcfactor must add cache to calcfactor originally posted by dehann in
| 0
|
19,649
| 26,008,904,910
|
IssuesEvent
|
2022-12-20 22:30:44
|
CSE201-project/PaperFriend-desktop-app
|
https://api.github.com/repos/CSE201-project/PaperFriend-desktop-app
|
closed
|
Load from file/ save to file activities and friends
|
file processing
|
As you can see at the top of mainwindow.cpp, there are a few vectors that are going to be shared by many components of the app. Here are the two you have to focus on:
std::vector<Activity>vector_activities;
std::vector<Friend>vector_friends;
They represent all the activities resp. friends that the user can select when writing an entry.
- When the app is launched, you should load form a file activities.json and friends.json the activities and friends that are in these two vectors.
- When the app is closed, you have to save back the current content of these two vectors in friends.json and activities.json
|
1.0
|
Load from file/ save to file activities and friends - As you can see at the top of mainwindow.cpp, there are a few vectors that are going to be shared by many components of the app. Here are the two you have to focus on:
std::vector<Activity>vector_activities;
std::vector<Friend>vector_friends;
They represent all the activities resp. friends that the user can select when writing an entry.
- When the app is launched, you should load form a file activities.json and friends.json the activities and friends that are in these two vectors.
- When the app is closed, you have to save back the current content of these two vectors in friends.json and activities.json
|
process
|
load from file save to file activities and friends as you can see at the top of mainwindow cpp there are a few vectors that are going to be shared by many components of the app here are the two you have to focus on std vector vector activities std vector vector friends they represent all the activities resp friends that the user can select when writing an entry when the app is launched you should load form a file activities json and friends json the activities and friends that are in these two vectors when the app is closed you have to save back the current content of these two vectors in friends json and activities json
| 1
|
20,948
| 27,807,732,079
|
IssuesEvent
|
2023-03-17 21:54:56
|
cse442-at-ub/project_s23-iweatherify
|
https://api.github.com/repos/cse442-at-ub/project_s23-iweatherify
|
closed
|
Add button navigation functionality for the Menu
|
Processing Task Sprint 2
|
**Tests**
What the Menu looks like: 
*Test 1*
1) Go to https://github.com/cse442-at-ub/project_s23-iweatherify/tree/55_menu_button_nav
2) Click on <> Code
3) Download as a Zip
4) Follow the instructions in README.md on how to start and run on localhost
5) Observe that as a non-logged-in user on the homepage, the menu can be opened/closed by pressing the hamburger menu
![Screenshot 2023-03-15 at 11.01.21 PM.png]
*Test 2*
1) Follow Test 1 Up until (4)
2) Observe that as a logged-in user on the homepage, the menu can be opened/closed by pressing the hamburger menu
![Screenshot 2023-03-15 at 11.01.21 PM.png]
*Test 2*
1) Follow Test 1 Up until (4)
2) From the homepage, click on the "register" button, it should take you to the register page
3) Confirm that the register button on the current registration page is not functional and is grayed out.
4) Return to the Homepage by clicking on that button and confirm it takes you to the non-logged-in homepage
5) From the homepage, click on the "login" button, it should take you to the login page
6) Confirm that the login button on the current login page is not functional and is grayed out.
7) Return to the Homepage by clicking on that button and confirm it takes you to the non-logged-in homepage
*Test 3*
1) Follow Test 1 Up until (4)
2) From the Register Page, click on the "login" button, it should take you to the login page
3) From the Login Page, click on the "Register" button it should take you to the register page
*Test 4*
1) Follow Test 1 Up until (4)
2) Go to the login page using the button navigation
3) Enter the username "zal" and password "kal" and press Login
4) Confirm that The Homepage button is grayed out
5) Click on the "Unit Settings" button, it should take you to the Unit Settings Page.
6) Open the Menu Once More and Click on Homepage confirming it's working
7) From the Homepage click on "Temperature Settings" confirming that it takes you to the temperature settings page
8) Click on the Homepage to return
*Test 5*
1) Follow Test 4 up until (4)
2) Open the menu and click on the Unit Settings button to take you to the Unit Settings Page
3) From this page click on "Temperature Settings", it should take you to the appropriate page
4) From the "Temperature Settings" Page open the menu and click on "Unit Settings", it should take you to the unit settings page
*Test 6*
1) Follow Test 4 up until (4)
2) Click on the logout button, confirming it logs you out
3) Log back into your account with "zal" as the username "kal" as the password and login
4) Navigate to the Temperature Settings Page and click on log out from that page
5) Log back into your account with "zal" as the username "kal" as the password and login
6) Navigate to the Unit Settings Page and click on log out from that page
7) Confirm that Log Out Functions Properly

|
1.0
|
Add button navigation functionality for the Menu - **Tests**
What the Menu looks like: 
*Test 1*
1) Go to https://github.com/cse442-at-ub/project_s23-iweatherify/tree/55_menu_button_nav
2) Click on <> Code
3) Download as a Zip
4) Follow the instructions in README.md on how to start and run on localhost
5) Observe that as a non-logged-in user on the homepage, the menu can be opened/closed by pressing the hamburger menu
![Screenshot 2023-03-15 at 11.01.21 PM.png]
*Test 2*
1) Follow Test 1 Up until (4)
2) Observe that as a logged-in user on the homepage, the menu can be opened/closed by pressing the hamburger menu
![Screenshot 2023-03-15 at 11.01.21 PM.png]
*Test 2*
1) Follow Test 1 Up until (4)
2) From the homepage, click on the "register" button, it should take you to the register page
3) Confirm that the register button on the current registration page is not functional and is grayed out.
4) Return to the Homepage by clicking on that button and confirm it takes you to the non-logged-in homepage
5) From the homepage, click on the "login" button, it should take you to the login page
6) Confirm that the login button on the current login page is not functional and is grayed out.
7) Return to the Homepage by clicking on that button and confirm it takes you to the non-logged-in homepage
*Test 3*
1) Follow Test 1 Up until (4)
2) From the Register Page, click on the "login" button, it should take you to the login page
3) From the Login Page, click on the "Register" button it should take you to the register page
*Test 4*
1) Follow Test 1 Up until (4)
2) Go to the login page using the button navigation
3) Enter the username "zal" and password "kal" and press Login
4) Confirm that The Homepage button is grayed out
5) Click on the "Unit Settings" button, it should take you to the Unit Settings Page.
6) Open the Menu Once More and Click on Homepage confirming it's working
7) From the Homepage click on "Temperature Settings" confirming that it takes you to the temperature settings page
8) Click on the Homepage to return
*Test 5*
1) Follow Test 4 up until (4)
2) Open the menu and click on the Unit Settings button to take you to the Unit Settings Page
3) From this page click on "Temperature Settings", it should take you to the appropriate page
4) From the "Temperature Settings" Page open the menu and click on "Unit Settings", it should take you to the unit settings page
*Test 6*
1) Follow Test 4 up until (4)
2) Click on the logout button, confirming it logs you out
3) Log back into your account with "zal" as the username "kal" as the password and login
4) Navigate to the Temperature Settings Page and click on log out from that page
5) Log back into your account with "zal" as the username "kal" as the password and login
6) Navigate to the Unit Settings Page and click on log out from that page
7) Confirm that Log Out Functions Properly

|
process
|
add button navigation functionality for the menu tests what the menu looks like test go to click on code download as a zip follow the instructions in readme md on how to start and run on localhost observe that as a non logged in user on the homepage the menu can be opened closed by pressing the hamburger menu test follow test up until observe that as a logged in user on the homepage the menu can be opened closed by pressing the hamburger menu test follow test up until from the homepage click on the register button it should take you to the register page confirm that the register button on the current registration page is not functional and is grayed out return to the homepage by clicking on that button and confirm it takes you to the non logged in homepage from the homepage click on the login button it should take you to the login page confirm that the login button on the current login page is not functional and is grayed out return to the homepage by clicking on that button and confirm it takes you to the non logged in homepage test follow test up until from the register page click on the login button it should take you to the login page from the login page click on the register button it should take you to the register page test follow test up until go to the login page using the button navigation enter the username zal and password kal and press login confirm that the homepage button is grayed out click on the unit settings button it should take you to the unit settings page open the menu once more and click on homepage confirming it s working from the homepage click on temperature settings confirming that it takes you to the temperature settings page click on the homepage to return test follow test up until open the menu and click on the unit settings button to take you to the unit settings page from this page click on temperature settings it should take you to the appropriate page from the temperature settings page open the menu and click on unit settings it should take you to the unit settings page test follow test up until click on the logout button confirming it logs you out log back into your account with zal as the username kal as the password and login navigate to the temperature settings page and click on log out from that page log back into your account with zal as the username kal as the password and login navigate to the unit settings page and click on log out from that page confirm that log out functions properly
| 1
|
1,113
| 3,590,326,369
|
IssuesEvent
|
2016-02-01 04:43:19
|
dotnet/corefx
|
https://api.github.com/repos/dotnet/corefx
|
closed
|
ProcessTests.TestWorkingSet64 failed on OS X in CI
|
System.Diagnostics.Process
|
http://dotnet-ci.cloudapp.net/job/dotnet_corefx_mac_debug_tst/31/console
```
21:51:43 System.Diagnostics.ProcessTests.ProcessTests.TestWorkingSet64 [FAIL]
21:51:43 Assert+WrapperXunitException : File path: d:\j\workspace\dotnet_corefx_windows_debug\src\System.Diagnostics.Process\tests\ProcessTests.cs. Line: 339
21:51:43 ---- Assert.True() Failure
21:51:43 Expected: True
21:51:43 Actual: False
21:51:43 Stack Trace:
21:51:43 at Assert.WrapException(Exception e, String callerFilePath, Int32 callerLineNumber)
21:51:43 at Assert.True(Boolean condition, String userMessage, String path, Int32 line)
21:51:43 at System.Diagnostics.ProcessTests.ProcessTests.TestWorkingSet64()
21:51:43 ----- Inner Stack Trace -----
21:51:43 at Assert.True(Boolean condition, String userMessage, String path, Int32 line)
```
|
1.0
|
ProcessTests.TestWorkingSet64 failed on OS X in CI - http://dotnet-ci.cloudapp.net/job/dotnet_corefx_mac_debug_tst/31/console
```
21:51:43 System.Diagnostics.ProcessTests.ProcessTests.TestWorkingSet64 [FAIL]
21:51:43 Assert+WrapperXunitException : File path: d:\j\workspace\dotnet_corefx_windows_debug\src\System.Diagnostics.Process\tests\ProcessTests.cs. Line: 339
21:51:43 ---- Assert.True() Failure
21:51:43 Expected: True
21:51:43 Actual: False
21:51:43 Stack Trace:
21:51:43 at Assert.WrapException(Exception e, String callerFilePath, Int32 callerLineNumber)
21:51:43 at Assert.True(Boolean condition, String userMessage, String path, Int32 line)
21:51:43 at System.Diagnostics.ProcessTests.ProcessTests.TestWorkingSet64()
21:51:43 ----- Inner Stack Trace -----
21:51:43 at Assert.True(Boolean condition, String userMessage, String path, Int32 line)
```
|
process
|
processtests failed on os x in ci system diagnostics processtests processtests assert wrapperxunitexception file path d j workspace dotnet corefx windows debug src system diagnostics process tests processtests cs line assert true failure expected true actual false stack trace at assert wrapexception exception e string callerfilepath callerlinenumber at assert true boolean condition string usermessage string path line at system diagnostics processtests processtests inner stack trace at assert true boolean condition string usermessage string path line
| 1
|
380,132
| 26,403,897,084
|
IssuesEvent
|
2023-01-13 05:33:14
|
tcmetzger/sphinx-favicon
|
https://api.github.com/repos/tcmetzger/sphinx-favicon
|
closed
|
do you want to create a sphinx single page documentation ?
|
documentation enhancement
|
I saw sphinx in the dev-requirements.txt file and I was curious if you wanted to use it ?
|
1.0
|
do you want to create a sphinx single page documentation ? - I saw sphinx in the dev-requirements.txt file and I was curious if you wanted to use it ?
|
non_process
|
do you want to create a sphinx single page documentation i saw sphinx in the dev requirements txt file and i was curious if you wanted to use it
| 0
|
367,298
| 10,851,789,829
|
IssuesEvent
|
2019-11-13 11:28:47
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
www.google.com - desktop site instead of mobile site
|
browser-firefox-tablet engine-gecko priority-critical
|
<!-- @browser: Firefox Mobile (Tablet) 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Tablet; rv:68.0) Gecko/68.0 Firefox/68.0 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://www.google.com/#spf=1573546776401
**Browser / Version**: Firefox Mobile (Tablet) 68.0
**Operating System**: Android 8.1.0
**Tested Another Browser**: Yes
**Problem type**: Desktop site instead of mobile site
**Description**: when clicking on image in Google, site appears instead of photo
**Steps to Reproduce**:
[](https://webcompat.com/uploads/2019/11/9931b5f6-c80c-4174-90b2-519f2d2c0997.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20191017184803</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
www.google.com - desktop site instead of mobile site - <!-- @browser: Firefox Mobile (Tablet) 68.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Tablet; rv:68.0) Gecko/68.0 Firefox/68.0 -->
<!-- @reported_with: mobile-reporter -->
**URL**: https://www.google.com/#spf=1573546776401
**Browser / Version**: Firefox Mobile (Tablet) 68.0
**Operating System**: Android 8.1.0
**Tested Another Browser**: Yes
**Problem type**: Desktop site instead of mobile site
**Description**: when clicking on image in Google, site appears instead of photo
**Steps to Reproduce**:
[](https://webcompat.com/uploads/2019/11/9931b5f6-c80c-4174-90b2-519f2d2c0997.jpeg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20191017184803</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
desktop site instead of mobile site url browser version firefox mobile tablet operating system android tested another browser yes problem type desktop site instead of mobile site description when clicking on image in google site appears instead of photo steps to reproduce browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
| 0
|
19,529
| 25,840,079,329
|
IssuesEvent
|
2022-12-12 23:15:30
|
OpenDataScotland/the_od_bods
|
https://api.github.com/repos/OpenDataScotland/the_od_bods
|
opened
|
Extract category keywords from dataset title and description
|
data processing back end
|
**Is your feature request related to a problem? Please describe.**
Too many datasets are set as uncategorised. This is because current categorisation uses category keywords provided by the publisher. Where there are no keywords provided by the publisher, the dataset cannot be further categorised by us in the context of the ODS catalogue. There is a similar ticket #172 but it is a large ticket to tackle. This ticket is one step down, a subset just to extract keywords out of the dataset title and description to use for categorisation, categorisation still using the existing keyword matching system.
**Describe the solution you'd like**
- Combine dataset title and description into single string/ text body
- tokenise and remove stopwords
- for each remaining keyword, get frequency count in body and return a matching category
- retain top 5 most common categories (based on frequency counts) and set as dataset categories
**Describe alternatives you've considered**
- consider standardising casing and stemming for more accurate comparison
- consider that there may be no need to cap the number of categories - i.e. returning all categories may be appropriate. See this as a % of total categories - if most categories return most of the time, then it's a meaningless solution. But because we manually curate category keywords, it might actually be filtered enough.
- If the resulting categorisation takes too long (compared to current merge_data.py performance) then consider categorising on top n keywords only (instead of all keywords). The catch is that the top common words may not be useful keywords, but we may be satisfied enough with speed and small % of datasets left uncategorised.
- consider TF-IDF principles (although may be more appropriate at #172 stage)
**Additional context**
Completion of this ticket leaves #172 to be an exploratory piece using unsupervised learning, but still a step-up in performance until then.
|
1.0
|
Extract category keywords from dataset title and description - **Is your feature request related to a problem? Please describe.**
Too many datasets are set as uncategorised. This is because current categorisation uses category keywords provided by the publisher. Where there are no keywords provided by the publisher, the dataset cannot be further categorised by us in the context of the ODS catalogue. There is a similar ticket #172 but it is a large ticket to tackle. This ticket is one step down, a subset just to extract keywords out of the dataset title and description to use for categorisation, categorisation still using the existing keyword matching system.
**Describe the solution you'd like**
- Combine dataset title and description into single string/ text body
- tokenise and remove stopwords
- for each remaining keyword, get frequency count in body and return a matching category
- retain top 5 most common categories (based on frequency counts) and set as dataset categories
**Describe alternatives you've considered**
- consider standardising casing and stemming for more accurate comparison
- consider that there may be no need to cap the number of categories - i.e. returning all categories may be appropriate. See this as a % of total categories - if most categories return most of the time, then it's a meaningless solution. But because we manually curate category keywords, it might actually be filtered enough.
- If the resulting categorisation takes too long (compared to current merge_data.py performance) then consider categorising on top n keywords only (instead of all keywords). The catch is that the top common words may not be useful keywords, but we may be satisfied enough with speed and small % of datasets left uncategorised.
- consider TF-IDF principles (although may be more appropriate at #172 stage)
**Additional context**
Completion of this ticket leaves #172 to be an exploratory piece using unsupervised learning, but still a step-up in performance until then.
|
process
|
extract category keywords from dataset title and description is your feature request related to a problem please describe too many datasets are set as uncategorised this is because current categorisation uses category keywords provided by the publisher where there are no keywords provided by the publisher the dataset cannot be further categorised by us in the context of the ods catalogue there is a similar ticket but it is a large ticket to tackle this ticket is one step down a subset just to extract keywords out of the dataset title and description to use for categorisation categorisation still using the existing keyword matching system describe the solution you d like combine dataset title and description into single string text body tokenise and remove stopwords for each remaining keyword get frequency count in body and return a matching category retain top most common categories based on frequency counts and set as dataset categories describe alternatives you ve considered consider standardising casing and stemming for more accurate comparison consider that there may be no need to cap the number of categories i e returning all categories may be appropriate see this as a of total categories if most categories return most of the time then it s a meaningless solution but because we manually curate category keywords it might actually be filtered enough if the resulting categorisation takes too long compared to current merge data py performance then consider categorising on top n keywords only instead of all keywords the catch is that the top common words may not be useful keywords but we may be satisfied enough with speed and small of datasets left uncategorised consider tf idf principles although may be more appropriate at stage additional context completion of this ticket leaves to be an exploratory piece using unsupervised learning but still a step up in performance until then
| 1
|
10,136
| 13,044,162,426
|
IssuesEvent
|
2020-07-29 03:47:32
|
tikv/tikv
|
https://api.github.com/repos/tikv/tikv
|
closed
|
UCP: Migrate scalar function `JsonPrettySig` from TiDB
|
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
|
## Description
Port the scalar function `JsonPrettySig` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @breeswish
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
2.0
|
UCP: Migrate scalar function `JsonPrettySig` from TiDB -
## Description
Port the scalar function `JsonPrettySig` from TiDB to coprocessor.
## Score
* 50
## Mentor(s)
* @breeswish
## Recommended Skills
* Rust programming
## Learning Materials
Already implemented expressions ported from TiDB
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr)
- https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
|
process
|
ucp migrate scalar function jsonprettysig from tidb description port the scalar function jsonprettysig from tidb to coprocessor score mentor s breeswish recommended skills rust programming learning materials already implemented expressions ported from tidb
| 1
|
18,649
| 24,581,037,496
|
IssuesEvent
|
2022-10-13 15:37:42
|
GoogleCloudPlatform/fda-mystudies
|
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
|
closed
|
[FHIR] QuestionnaireResponse > JSON file > Response value is not getting mapped in the following scenario
|
Bug P0 Response datastore Process: Fixed Process: Tested dev
|
Steps:
1. In SB create a study with Text choice response type activity by using 'Other' answer options
2. launch the study
3. In mobile, Sign up or sign in to the mobile app
4. Enroll to the above study
5. Submit the response for above created activity by selecting Other option
6. Observe
AR: QuestionnaireResponse > JSON file > Response value is not getting mapped
ER: QuestionnaireResponse > JSON file > Response value should get mapped into JSON even though when the participant selects the Other option

|
2.0
|
[FHIR] QuestionnaireResponse > JSON file > Response value is not getting mapped in the following scenario - Steps:
1. In SB create a study with Text choice response type activity by using 'Other' answer options
2. launch the study
3. In mobile, Sign up or sign in to the mobile app
4. Enroll to the above study
5. Submit the response for above created activity by selecting Other option
6. Observe
AR: QuestionnaireResponse > JSON file > Response value is not getting mapped
ER: QuestionnaireResponse > JSON file > Response value should get mapped into JSON even though when the participant selects the Other option

|
process
|
questionnaireresponse json file response value is not getting mapped in the following scenario steps in sb create a study with text choice response type activity by using other answer options launch the study in mobile sign up or sign in to the mobile app enroll to the above study submit the response for above created activity by selecting other option observe ar questionnaireresponse json file response value is not getting mapped er questionnaireresponse json file response value should get mapped into json even though when the participant selects the other option
| 1
|
29,583
| 14,189,826,347
|
IssuesEvent
|
2020-11-14 02:40:44
|
timotheecour/Nim
|
https://api.github.com/repos/timotheecour/Nim
|
closed
|
--gc:arc 10_000 times slower
|
performance
|
from https://github.com/nim-lang/Nim/pull/15903#discussion_r523346150
nim r -d:danger -d:case1 --hints:off $timn_D/tests/nim/all/t11283.nim
OK! isNumeric3() time consuming:1.999999999999832e-06
nim r --gc:arc -d:danger -d:case1 --hints:off $timn_D/tests/nim/all/t11283.nim
OK! isNumeric3() time consuming:0.010939
```nim
import times
func isNumeric3*(s: string, enableNaNInf, enableLooseDot = false): bool =
## Checks whether the string is numeric.
## When the string is an integer, float or exponential, it returns true,
## otherwise it returns false.
## The `enableNaNInf` value indicates whether the `NaN` and `Inf` values
## are valid.
## The `enableLooseDot` value indicates whether loose point values such as
## `.9` and `9.` are valid.
##
## **Note:** The reason that `parseFloat()` is not used to achieve this is its
## poor performance.
let length = s.len
if length == 3:
if (s[0] in {'i', 'I'} and s[1] in {'n', 'N'} and s[2] in {'f', 'F'}) or
(s[0] in {'n', 'N'} and s[1] in {'a', 'A'} and s[2] in {'n', 'N'}):
if enableNaNInf:
return true
else:
return false
if length == 4:
if (s[0] in {'+', '-'} and s[1] in {'i', 'I'} and s[2] in {'n', 'N'} and s[3] in {'f', 'F'}) or
(s[0] in {'+', '-'} and s[1] in {'n', 'N'} and s[2] in {'a', 'A'} and s[3] in {'n', 'N'}):
if enableNaNInf:
return true
else:
return false
var eLeft, eRight, dot, e, num = false
let sHigh = s.len - 1
for i, si in s:
case si
of '+', '-':
if i == sHigh:
return false
if e == false:
if num:
return false
if eLeft:
return false
eLeft = true
else:
if num:
return false
if eRight:
return false
eRight = true
of '.':
if dot:
return false
if not enableLooseDot:
if num == false:
return false
else:
if i == sHigh:
return false
if s[i+1] in {'e', 'E'}:
return false
num = false
dot = true
of 'e', 'E':
if i == sHigh:
return false
if num or dot:
if e:
return false
num = false
else:
return false
e = true
of '0'..'9':
num = true
of '_':
if num == false:
return false
if dot and num == false:
return false
if e and num == false:
return false
else:
return false
return true
proc testIsNumeric3() =
var a = 0
let execptedValue = 160000000
let t0 = cputime()
for i in 0..9999999:
if isNumeric3("123"): a += 1
if isNumeric3("123.45"): a += 1
if isNumeric3("+123.45"): a += 1
if isNumeric3("123.45e-2"): a += 1
if isNumeric3("+123.45E-2"): a += 1
if isNumeric3("-123.45e2"): a += 1
if not isNumeric3("e123.45"): a += 1
if not isNumeric3("abc"): a += 1
if not isNumeric3("123abc"): a += 1
if not isNumeric3("123.45.6"): a += 1
if not isNumeric3("123.45e++5"): a += 1
if not isNumeric3("5.2+e1"): a += 1
if not isNumeric3(".9"): a += 1
if not isNumeric3("Inf"): a += 1
if not isNumeric3("-Inf"): a += 1
if not isNumeric3("NaN"): a += 1
let t1 = cputime() - t0
if a == execptedValue:
echo "OK! isNumeric3() time consuming:", t1
testIsNumeric3()
```
|
True
|
--gc:arc 10_000 times slower - from https://github.com/nim-lang/Nim/pull/15903#discussion_r523346150
nim r -d:danger -d:case1 --hints:off $timn_D/tests/nim/all/t11283.nim
OK! isNumeric3() time consuming:1.999999999999832e-06
nim r --gc:arc -d:danger -d:case1 --hints:off $timn_D/tests/nim/all/t11283.nim
OK! isNumeric3() time consuming:0.010939
```nim
import times
func isNumeric3*(s: string, enableNaNInf, enableLooseDot = false): bool =
## Checks whether the string is numeric.
## When the string is an integer, float or exponential, it returns true,
## otherwise it returns false.
## The `enableNaNInf` value indicates whether the `NaN` and `Inf` values
## are valid.
## The `enableLooseDot` value indicates whether loose point values such as
## `.9` and `9.` are valid.
##
## **Note:** The reason that `parseFloat()` is not used to achieve this is its
## poor performance.
let length = s.len
if length == 3:
if (s[0] in {'i', 'I'} and s[1] in {'n', 'N'} and s[2] in {'f', 'F'}) or
(s[0] in {'n', 'N'} and s[1] in {'a', 'A'} and s[2] in {'n', 'N'}):
if enableNaNInf:
return true
else:
return false
if length == 4:
if (s[0] in {'+', '-'} and s[1] in {'i', 'I'} and s[2] in {'n', 'N'} and s[3] in {'f', 'F'}) or
(s[0] in {'+', '-'} and s[1] in {'n', 'N'} and s[2] in {'a', 'A'} and s[3] in {'n', 'N'}):
if enableNaNInf:
return true
else:
return false
var eLeft, eRight, dot, e, num = false
let sHigh = s.len - 1
for i, si in s:
case si
of '+', '-':
if i == sHigh:
return false
if e == false:
if num:
return false
if eLeft:
return false
eLeft = true
else:
if num:
return false
if eRight:
return false
eRight = true
of '.':
if dot:
return false
if not enableLooseDot:
if num == false:
return false
else:
if i == sHigh:
return false
if s[i+1] in {'e', 'E'}:
return false
num = false
dot = true
of 'e', 'E':
if i == sHigh:
return false
if num or dot:
if e:
return false
num = false
else:
return false
e = true
of '0'..'9':
num = true
of '_':
if num == false:
return false
if dot and num == false:
return false
if e and num == false:
return false
else:
return false
return true
proc testIsNumeric3() =
var a = 0
let execptedValue = 160000000
let t0 = cputime()
for i in 0..9999999:
if isNumeric3("123"): a += 1
if isNumeric3("123.45"): a += 1
if isNumeric3("+123.45"): a += 1
if isNumeric3("123.45e-2"): a += 1
if isNumeric3("+123.45E-2"): a += 1
if isNumeric3("-123.45e2"): a += 1
if not isNumeric3("e123.45"): a += 1
if not isNumeric3("abc"): a += 1
if not isNumeric3("123abc"): a += 1
if not isNumeric3("123.45.6"): a += 1
if not isNumeric3("123.45e++5"): a += 1
if not isNumeric3("5.2+e1"): a += 1
if not isNumeric3(".9"): a += 1
if not isNumeric3("Inf"): a += 1
if not isNumeric3("-Inf"): a += 1
if not isNumeric3("NaN"): a += 1
let t1 = cputime() - t0
if a == execptedValue:
echo "OK! isNumeric3() time consuming:", t1
testIsNumeric3()
```
|
non_process
|
gc arc times slower from nim r d danger d hints off timn d tests nim all nim ok time consuming nim r gc arc d danger d hints off timn d tests nim all nim ok time consuming nim import times func s string enablenaninf enableloosedot false bool checks whether the string is numeric when the string is an integer float or exponential it returns true otherwise it returns false the enablenaninf value indicates whether the nan and inf values are valid the enableloosedot value indicates whether loose point values such as and are valid note the reason that parsefloat is not used to achieve this is its poor performance let length s len if length if s in i i and s in n n and s in f f or s in n n and s in a a and s in n n if enablenaninf return true else return false if length if s in and s in i i and s in n n and s in f f or s in and s in n n and s in a a and s in n n if enablenaninf return true else return false var eleft eright dot e num false let shigh s len for i si in s case si of if i shigh return false if e false if num return false if eleft return false eleft true else if num return false if eright return false eright true of if dot return false if not enableloosedot if num false return false else if i shigh return false if s in e e return false num false dot true of e e if i shigh return false if num or dot if e return false num false else return false e true of num true of if num false return false if dot and num false return false if e and num false return false else return false return true proc var a let execptedvalue let cputime for i in if a if a if a if a if a if a if not a if not abc a if not a if not a if not a if not a if not a if not inf a if not inf a if not nan a let cputime if a execptedvalue echo ok time consuming
| 0
|
4,334
| 7,242,199,103
|
IssuesEvent
|
2018-02-14 06:16:38
|
muflihun/residue
|
https://api.github.com/repos/muflihun/residue
|
closed
|
Deadlock detected while dispatching log
|
area: log-processing edge-case type: bug
|
```
19:03:02,606 [LogHandler] [vTRACE] [session.cc:86] Adding bytes
19:03:02,608 [LogDispatcher] [vDEBUG] [log-request-handler.cc:88] Pausing schedule for client integrity
19:03:02,609 [LogHandler] [vTRACE] [session.cc:75] Received: 522 bytes
19:03:02,609 [LogHandler] [vTRACE] [session.cc:102] Read bytes: [..snip] [size: 522]
19:03:02,609 [LogHandler] [vDEBUG] [session.cc:171] Sending {r:0}
19:03:02,609 [LogHandler] [vTRACE] [session.cc:86] Adding bytes
19:03:02,609 [LogDispatcher] [vCRAZY] [log-request-handler.cc:100] -----============= [ BEGIN ] =============-----
19:03:02,610 [LogDispatcher] [vDEBUG] [request-handler.h:144] Raw request: [..snip]
19:03:02,610 [LogDispatcher] [vDEBUG] [request-handler.cc:58] Client: [..snip]
19:03:02,610 [LogDispatcher] [vCRAZY] [request-handler.cc:59] IV: [..snip]
19:03:02,610 [LogDispatcher] [vCRAZY] [request-handler.cc:66] Data (base64): [..snip]
19:03:02,610 [LogDispatcher] [vCRAZY] [request-handler.cc:69] Ripe command: [..snip]
19:03:02,610 [LogDispatcher] [vCRAZY] [request-handler.cc:76] Plain request: [..snip]
19:03:02,610 [LogDispatcher] [vTRACE] [request-handler.h:148] Decryption finished (b64): [..snip]
19:03:02,610 [LogDispatcher] [vTRACE] [request-handler.h:159] Decompressing: [..snip]
19:03:02,611 [LogDispatcher] [vTRACE] [request-handler.h:170] Decompression finished (raw): [{... (removed)}]
19:03:02,612 [LogDispatcher] [vDEBUG] [log-request-handler.cc:115] Is bulk? true
19:03:02,612 [LogDispatcher] [vDEBUG] [log-request-handler.cc:124] Request client: 0x10065cfa8
19:03:02,612 [LogHandler] [vTRACE] [session.cc:75] Received: 522 bytes
19:03:02,612 [LogDispatcher] [vDEBUG] [log-request-handler.cc:206] Force check: true, clientRef: 0x70000f9729a0, *clientRef: muflihun00102030, bypassChecks: false
19:03:02,612 [LogHandler] [vTRACE] [session.cc:102] Read bytes: 29277885c8e3830c20fb91fab830194b:muflihun00102030:zaAqRKANlz9SElDUXg6ePbeykddpAyJjxC3TEL/yutU5HIryqLpN1xmsEncDHizpniIeaCIY7kfUjuzTaXwYJC0dxHxlssq5Lm7PzgPCyyxa7POh0quW1zooTSlBMvUiuCWamAdfzz3WPPjjbMSyvqtgMmbFIhs4qMMMt3kKZbli8kk9Hmv5XV8Ltz+/tLX4ppiOn6qSFh/pIQ0LNaBVK0wHXLfGO+AN58vc1D50JT8vA00XUjkzNNmeasekIntZDoRut/WhgqVijoUsT27sY9WzrsdaLuIJIAUhtewqvAGgyeplI+lucgHWtWKVR+HQQ887giQwW6+80FrU0yIZ/yEhDnFbCul3Zc9L3vuuBHvrAgjYvDPkA4MZPwBkFWSh6anUbJ5zsNgEnMP2tdDI+ZLnSVfvQ+ZNEPeVXYE15la/XVPoe0/b/2MqY3HPUrYEtTJONqT/XrZKl1HXbqxp4w== [size: 522]
19:03:02,613 [LogDispatcher] [vDEBUG] [client.cc:65] Checking token 37953058 (client [muflihun00102030])
```
|
1.0
|
Deadlock detected while dispatching log - ```
19:03:02,606 [LogHandler] [vTRACE] [session.cc:86] Adding bytes
19:03:02,608 [LogDispatcher] [vDEBUG] [log-request-handler.cc:88] Pausing schedule for client integrity
19:03:02,609 [LogHandler] [vTRACE] [session.cc:75] Received: 522 bytes
19:03:02,609 [LogHandler] [vTRACE] [session.cc:102] Read bytes: [..snip] [size: 522]
19:03:02,609 [LogHandler] [vDEBUG] [session.cc:171] Sending {r:0}
19:03:02,609 [LogHandler] [vTRACE] [session.cc:86] Adding bytes
19:03:02,609 [LogDispatcher] [vCRAZY] [log-request-handler.cc:100] -----============= [ BEGIN ] =============-----
19:03:02,610 [LogDispatcher] [vDEBUG] [request-handler.h:144] Raw request: [..snip]
19:03:02,610 [LogDispatcher] [vDEBUG] [request-handler.cc:58] Client: [..snip]
19:03:02,610 [LogDispatcher] [vCRAZY] [request-handler.cc:59] IV: [..snip]
19:03:02,610 [LogDispatcher] [vCRAZY] [request-handler.cc:66] Data (base64): [..snip]
19:03:02,610 [LogDispatcher] [vCRAZY] [request-handler.cc:69] Ripe command: [..snip]
19:03:02,610 [LogDispatcher] [vCRAZY] [request-handler.cc:76] Plain request: [..snip]
19:03:02,610 [LogDispatcher] [vTRACE] [request-handler.h:148] Decryption finished (b64): [..snip]
19:03:02,610 [LogDispatcher] [vTRACE] [request-handler.h:159] Decompressing: [..snip]
19:03:02,611 [LogDispatcher] [vTRACE] [request-handler.h:170] Decompression finished (raw): [{... (removed)}]
19:03:02,612 [LogDispatcher] [vDEBUG] [log-request-handler.cc:115] Is bulk? true
19:03:02,612 [LogDispatcher] [vDEBUG] [log-request-handler.cc:124] Request client: 0x10065cfa8
19:03:02,612 [LogHandler] [vTRACE] [session.cc:75] Received: 522 bytes
19:03:02,612 [LogDispatcher] [vDEBUG] [log-request-handler.cc:206] Force check: true, clientRef: 0x70000f9729a0, *clientRef: muflihun00102030, bypassChecks: false
19:03:02,612 [LogHandler] [vTRACE] [session.cc:102] Read bytes: 29277885c8e3830c20fb91fab830194b:muflihun00102030:zaAqRKANlz9SElDUXg6ePbeykddpAyJjxC3TEL/yutU5HIryqLpN1xmsEncDHizpniIeaCIY7kfUjuzTaXwYJC0dxHxlssq5Lm7PzgPCyyxa7POh0quW1zooTSlBMvUiuCWamAdfzz3WPPjjbMSyvqtgMmbFIhs4qMMMt3kKZbli8kk9Hmv5XV8Ltz+/tLX4ppiOn6qSFh/pIQ0LNaBVK0wHXLfGO+AN58vc1D50JT8vA00XUjkzNNmeasekIntZDoRut/WhgqVijoUsT27sY9WzrsdaLuIJIAUhtewqvAGgyeplI+lucgHWtWKVR+HQQ887giQwW6+80FrU0yIZ/yEhDnFbCul3Zc9L3vuuBHvrAgjYvDPkA4MZPwBkFWSh6anUbJ5zsNgEnMP2tdDI+ZLnSVfvQ+ZNEPeVXYE15la/XVPoe0/b/2MqY3HPUrYEtTJONqT/XrZKl1HXbqxp4w== [size: 522]
19:03:02,613 [LogDispatcher] [vDEBUG] [client.cc:65] Checking token 37953058 (client [muflihun00102030])
```
|
process
|
deadlock detected while dispatching log adding bytes pausing schedule for client integrity received bytes read bytes sending r adding bytes raw request client iv data ripe command plain request decryption finished decompressing decompression finished raw is bulk true request client received bytes force check true clientref clientref bypasschecks false read bytes lucghwtwkvr zlnsvfvq b checking token client
| 1
|
4,580
| 2,734,424,272
|
IssuesEvent
|
2015-04-17 19:49:15
|
uonline/universe
|
https://api.github.com/repos/uonline/universe
|
opened
|
2.0
|
design discussion
|
Сижу вот, рисую дизайны. Придумал такой:

Очень нравится, в два раза удобнее нынешнего. Только непонятно, кто нам столько картинок отрисует.
|
1.0
|
2.0 - Сижу вот, рисую дизайны. Придумал такой:

Очень нравится, в два раза удобнее нынешнего. Только непонятно, кто нам столько картинок отрисует.
|
non_process
|
сижу вот рисую дизайны придумал такой очень нравится в два раза удобнее нынешнего только непонятно кто нам столько картинок отрисует
| 0
|
37,633
| 8,337,011,311
|
IssuesEvent
|
2018-09-28 09:40:14
|
Yoast/YoastSEO.js
|
https://api.github.com/repos/Yoast/YoastSEO.js
|
closed
|
Refactor SubheadingsKeyword assessment
|
code-review lingo morpho-syno
|
Refactor the subheadingsKeywordAssessment to follow the following schema:
- A subheading is considered to reflect the topic if > half of content words from the keyphrase are used in it.
- GOOD if 30-75% of subheadings reflect the topic,
- BAD otherwise.
Using synonyms: yes
|
1.0
|
Refactor SubheadingsKeyword assessment - Refactor the subheadingsKeywordAssessment to follow the following schema:
- A subheading is considered to reflect the topic if > half of content words from the keyphrase are used in it.
- GOOD if 30-75% of subheadings reflect the topic,
- BAD otherwise.
Using synonyms: yes
|
non_process
|
refactor subheadingskeyword assessment refactor the subheadingskeywordassessment to follow the following schema a subheading is considered to reflect the topic if half of content words from the keyphrase are used in it good if of subheadings reflect the topic bad otherwise using synonyms yes
| 0
|
459,927
| 13,201,501,732
|
IssuesEvent
|
2020-08-14 10:16:21
|
strapi/strapi
|
https://api.github.com/repos/strapi/strapi
|
closed
|
Setting a datetime in a component as an entry title causes a react error when setting the datetime
|
priority: low status: can't reproduce
|
**Describe the bug**
Setting a datetime's value inside of a component using the calendar UI causes a react error if the datetime is the entry title for the component.
**Steps to reproduce the behavior**
(Using the admin panel UI)
1. Create a content type
2. Create a piece of content using that type
3. Create a component.
4. Add a datetime to the component
5. Go to the content created in step 2
6. Using the "edit the view" button, edit the component's layout
7. Change the entry title to the datetime added in step 4
8. Go back to edit the content created in step 2
9. In the component, attempt to set the dateTime
10. React error: https://imgur.com/a/9WYX3UY
**Expected behavior**
The date is set and the entry title updates accordingly.
Node.js version: 12.14.1
NPM version: 6.13.4
Strapi version: Strapi v3.0.0
Database: Postgres
Operating system: ubuntu 18.04.04 lts
|
1.0
|
Setting a datetime in a component as an entry title causes a react error when setting the datetime - **Describe the bug**
Setting a datetime's value inside of a component using the calendar UI causes a react error if the datetime is the entry title for the component.
**Steps to reproduce the behavior**
(Using the admin panel UI)
1. Create a content type
2. Create a piece of content using that type
3. Create a component.
4. Add a datetime to the component
5. Go to the content created in step 2
6. Using the "edit the view" button, edit the component's layout
7. Change the entry title to the datetime added in step 4
8. Go back to edit the content created in step 2
9. In the component, attempt to set the dateTime
10. React error: https://imgur.com/a/9WYX3UY
**Expected behavior**
The date is set and the entry title updates accordingly.
Node.js version: 12.14.1
NPM version: 6.13.4
Strapi version: Strapi v3.0.0
Database: Postgres
Operating system: ubuntu 18.04.04 lts
|
non_process
|
setting a datetime in a component as an entry title causes a react error when setting the datetime describe the bug setting a datetime s value inside of a component using the calendar ui causes a react error if the datetime is the entry title for the component steps to reproduce the behavior using the admin panel ui create a content type create a piece of content using that type create a component add a datetime to the component go to the content created in step using the edit the view button edit the component s layout change the entry title to the datetime added in step go back to edit the content created in step in the component attempt to set the datetime react error expected behavior the date is set and the entry title updates accordingly node js version npm version strapi version strapi database postgres operating system ubuntu lts
| 0
|
10,948
| 13,756,464,341
|
IssuesEvent
|
2020-10-06 19:58:51
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Freeze and crash while trying to rename condition in conditional branch
|
Bug Modeller Processing
|
I have a Conditional branch in Model and other algorithm depends on it. If I try to rename condition branch -> Qgis freezes and crashes.

This can be avoided though if I first turn off dependency, then rename my condition branch and turn on dependency again

Sample project with model attached
[Bug with conditional branch.zip](https://github.com/qgis/QGIS/files/5287404/Bug.with.conditional.branch.zip)
P.s. Windows 10, Qgis 3.14 and Master
|
1.0
|
Freeze and crash while trying to rename condition in conditional branch - I have a Conditional branch in Model and other algorithm depends on it. If I try to rename condition branch -> Qgis freezes and crashes.

This can be avoided though if I first turn off dependency, then rename my condition branch and turn on dependency again

Sample project with model attached
[Bug with conditional branch.zip](https://github.com/qgis/QGIS/files/5287404/Bug.with.conditional.branch.zip)
P.s. Windows 10, Qgis 3.14 and Master
|
process
|
freeze and crash while trying to rename condition in conditional branch i have a conditional branch in model and other algorithm depends on it if i try to rename condition branch qgis freezes and crashes this can be avoided though if i first turn off dependency then rename my condition branch and turn on dependency again sample project with model attached p s windows qgis and master
| 1
|
9,074
| 12,147,383,330
|
IssuesEvent
|
2020-04-24 12:56:31
|
GoogleCloudPlatform/dotnet-docs-samples
|
https://api.github.com/repos/GoogleCloudPlatform/dotnet-docs-samples
|
closed
|
Logging: TestListEntries is timing out.
|
api: logging priority: p1 type: process
|
Even after 10 minutes. I've skipped it in #1001 .
I'll look at it.
|
1.0
|
Logging: TestListEntries is timing out. - Even after 10 minutes. I've skipped it in #1001 .
I'll look at it.
|
process
|
logging testlistentries is timing out even after minutes i ve skipped it in i ll look at it
| 1
|
19,645
| 26,006,104,451
|
IssuesEvent
|
2022-12-20 19:32:09
|
openxla/stablehlo
|
https://api.github.com/repos/openxla/stablehlo
|
reopened
|
Nonconforming syntax in multiple markdown files
|
Process
|
### What happened?
Although some MD parsers are pretty forgiving (including GitHub's), the current markdown files in stablehlo include a variety of syntax errors that do not conform to the basic [MD syntax rules](https://daringfireball.net/projects/markdown/syntax). Examples include improper indentation for lists, unescaped asterisks signs (meant for bold or italic text), and missing blank lines between headings and paragraphs.
We should follow traditional MD syntax rules to ensure well-structured files that are easy to read, use an agreed-upon structure that can be tested, and that are compatible with more than one MD parser.
The simplest solution is to [markdownlint](https://github.com/markdownlint/markdownlint) and fix various syntax and style issues.
We can submit fixes for individual files, and make exceptions as appropriate because this tool allows you to disable certain rules in the linter. Then, once the files are cleaned up, we should enable a GitHub Action to run markdownlint on all new pull requests with .md files so we can avoid re-introduction of syntax errors.
### Steps to reproduce your issue
_No response_
### Version information
_No response_
|
1.0
|
Nonconforming syntax in multiple markdown files - ### What happened?
Although some MD parsers are pretty forgiving (including GitHub's), the current markdown files in stablehlo include a variety of syntax errors that do not conform to the basic [MD syntax rules](https://daringfireball.net/projects/markdown/syntax). Examples include improper indentation for lists, unescaped asterisks signs (meant for bold or italic text), and missing blank lines between headings and paragraphs.
We should follow traditional MD syntax rules to ensure well-structured files that are easy to read, use an agreed-upon structure that can be tested, and that are compatible with more than one MD parser.
The simplest solution is to [markdownlint](https://github.com/markdownlint/markdownlint) and fix various syntax and style issues.
We can submit fixes for individual files, and make exceptions as appropriate because this tool allows you to disable certain rules in the linter. Then, once the files are cleaned up, we should enable a GitHub Action to run markdownlint on all new pull requests with .md files so we can avoid re-introduction of syntax errors.
### Steps to reproduce your issue
_No response_
### Version information
_No response_
|
process
|
nonconforming syntax in multiple markdown files what happened although some md parsers are pretty forgiving including github s the current markdown files in stablehlo include a variety of syntax errors that do not conform to the basic examples include improper indentation for lists unescaped asterisks signs meant for bold or italic text and missing blank lines between headings and paragraphs we should follow traditional md syntax rules to ensure well structured files that are easy to read use an agreed upon structure that can be tested and that are compatible with more than one md parser the simplest solution is to and fix various syntax and style issues we can submit fixes for individual files and make exceptions as appropriate because this tool allows you to disable certain rules in the linter then once the files are cleaned up we should enable a github action to run markdownlint on all new pull requests with md files so we can avoid re introduction of syntax errors steps to reproduce your issue no response version information no response
| 1
|
471,208
| 13,562,508,671
|
IssuesEvent
|
2020-09-18 07:00:21
|
qlcchain/qlc-hub
|
https://api.github.com/repos/qlcchain/qlc-hub
|
closed
|
release v1.0
|
Priority: High Type: Maintenance
|
- [x] prepare NEO/ETH signer accounts @ariesunny
- [x] update deploy scripts @gythialy
- [x] tag v1.0 @zengchen221
- [ ] deploy QLC staking contract @zengchen221
|
1.0
|
release v1.0 - - [x] prepare NEO/ETH signer accounts @ariesunny
- [x] update deploy scripts @gythialy
- [x] tag v1.0 @zengchen221
- [ ] deploy QLC staking contract @zengchen221
|
non_process
|
release prepare neo eth signer accounts ariesunny update deploy scripts gythialy tag deploy qlc staking contract
| 0
|
16,779
| 21,963,281,473
|
IssuesEvent
|
2022-05-24 17:37:05
|
cypress-io/cypress
|
https://api.github.com/repos/cypress-io/cypress
|
closed
|
How to use with customize-cra
|
type: question npm: @cypress/webpack-preprocessor
|
<!-- Want a bug fixed quickly? Please provide a repository to reproduce the issue. -->
How to use with `customize-cra`
package.json
```json
{
"name": "package",
"version": "1.3.0",
"private": true,
"scripts": {
"start": "react-native start",
"android": "react-native run-android",
"ios": "react-native run-ios",
"web": "REACT_APP_ENV=development react-app-rewired start",
"build": "echo Building to ${REACT_APP_ENV} environment && react-app-rewired build",
"build-dev": "REACT_APP_ENV=development CI=false npm run build",
"build-prod": "REACT_APP_ENV=production CI=true npm run build",
"build-android": "cd android && ./gradlew bundleRelease",
"test": "jest --passWithNoTests",
"cy:open": "NODE_ENV=test cypress open",
"cy:run:all": "NODE_ENV=test cypress run --config video=false",
"cy:run:smoke": "NODE_ENV=test cypress run --config video=false,integrationFolder=cypress/integration/smoke",
"serve": "serve -s build -p 3000",
"serve-test-cypress": "start-test serve 3000 cy:run:all"
},
"dependencies": {
"react": "16.9.0",
"react-native": "0.61.5",
"react-native-vector-icons": "^4.6.0",
"react-native-web": "^0.11.7",
"react-scripts": "^3.4.1",
...
},
"devDependencies": {
"@babel/core": "^7.9.0",
"@babel/plugin-proposal-class-properties": "^7.8.3",
"@babel/runtime": "^7.9.2",
"@cypress/webpack-preprocessor": "^4.1.3",
"babel-jest": "^24.9.0",
"babel-plugin-react-native-web": "^0.12.2",
"customize-cra": "^0.9.1",
"cypress": "^4.3.0",
"husky": "^4.2.3",
"jest": "^24.9.0",
"react-app-rewired": "^2.1.5",
"react-test-renderer": "16.9.0",
"serve": "^11.3.0",
"start-server-and-test": "^1.10.11",
...
},
"jest": {
"preset": "react-native",
"collectCoverageFrom": [
"src/**/*.{js,jsx,mjs}"
],
"testEnvironment": "node",
"testURL": "http://localhost",
"moduleNameMapper": {
"^react-native$": "react-native-web"
},
},
"husky": {
"hooks": {
"pre-commit": "npm run lint"
}
},
"browserslist": {
"production": [
">0.2%",
"not dead",
"not op_mini all"
],
"development": [
"last 1 chrome version",
"last 1 firefox version",
"last 1 safari version"
]
}
}
```
config-overrides.js
```javascript
const {
override,
disableEsLint,
babelInclude,
addBabelPlugin,
addExternalBabelPlugins,
addWebpackPlugin
} = require("customize-cra");
const webpack = require("webpack");
const path = require("path");
module.exports = override(
disableEsLint(),
addBabelPlugin("react-native-web"),
addExternalBabelPlugins("@babel/plugin-proposal-class-properties"),
babelInclude([
path.resolve("src"),
path.resolve("node_modules/react-native-vector-icons")
]),
addWebpackPlugin(
new webpack.DefinePlugin({
__DEV__: process.env.REACT_APP_ENV === "development" || process.env.NODE_ENV === 'development'
})
)
);
```
- Operating System: Ubuntu 19.04
- Cypress Version: 4.3.0
- Browser Version:
|
1.0
|
How to use with customize-cra - <!-- Want a bug fixed quickly? Please provide a repository to reproduce the issue. -->
How to use with `customize-cra`
package.json
```json
{
"name": "package",
"version": "1.3.0",
"private": true,
"scripts": {
"start": "react-native start",
"android": "react-native run-android",
"ios": "react-native run-ios",
"web": "REACT_APP_ENV=development react-app-rewired start",
"build": "echo Building to ${REACT_APP_ENV} environment && react-app-rewired build",
"build-dev": "REACT_APP_ENV=development CI=false npm run build",
"build-prod": "REACT_APP_ENV=production CI=true npm run build",
"build-android": "cd android && ./gradlew bundleRelease",
"test": "jest --passWithNoTests",
"cy:open": "NODE_ENV=test cypress open",
"cy:run:all": "NODE_ENV=test cypress run --config video=false",
"cy:run:smoke": "NODE_ENV=test cypress run --config video=false,integrationFolder=cypress/integration/smoke",
"serve": "serve -s build -p 3000",
"serve-test-cypress": "start-test serve 3000 cy:run:all"
},
"dependencies": {
"react": "16.9.0",
"react-native": "0.61.5",
"react-native-vector-icons": "^4.6.0",
"react-native-web": "^0.11.7",
"react-scripts": "^3.4.1",
...
},
"devDependencies": {
"@babel/core": "^7.9.0",
"@babel/plugin-proposal-class-properties": "^7.8.3",
"@babel/runtime": "^7.9.2",
"@cypress/webpack-preprocessor": "^4.1.3",
"babel-jest": "^24.9.0",
"babel-plugin-react-native-web": "^0.12.2",
"customize-cra": "^0.9.1",
"cypress": "^4.3.0",
"husky": "^4.2.3",
"jest": "^24.9.0",
"react-app-rewired": "^2.1.5",
"react-test-renderer": "16.9.0",
"serve": "^11.3.0",
"start-server-and-test": "^1.10.11",
...
},
"jest": {
"preset": "react-native",
"collectCoverageFrom": [
"src/**/*.{js,jsx,mjs}"
],
"testEnvironment": "node",
"testURL": "http://localhost",
"moduleNameMapper": {
"^react-native$": "react-native-web"
},
},
"husky": {
"hooks": {
"pre-commit": "npm run lint"
}
},
"browserslist": {
"production": [
">0.2%",
"not dead",
"not op_mini all"
],
"development": [
"last 1 chrome version",
"last 1 firefox version",
"last 1 safari version"
]
}
}
```
config-overrides.js
```javascript
const {
override,
disableEsLint,
babelInclude,
addBabelPlugin,
addExternalBabelPlugins,
addWebpackPlugin
} = require("customize-cra");
const webpack = require("webpack");
const path = require("path");
module.exports = override(
disableEsLint(),
addBabelPlugin("react-native-web"),
addExternalBabelPlugins("@babel/plugin-proposal-class-properties"),
babelInclude([
path.resolve("src"),
path.resolve("node_modules/react-native-vector-icons")
]),
addWebpackPlugin(
new webpack.DefinePlugin({
__DEV__: process.env.REACT_APP_ENV === "development" || process.env.NODE_ENV === 'development'
})
)
);
```
- Operating System: Ubuntu 19.04
- Cypress Version: 4.3.0
- Browser Version:
|
process
|
how to use with customize cra how to use with customize cra package json json name package version private true scripts start react native start android react native run android ios react native run ios web react app env development react app rewired start build echo building to react app env environment react app rewired build build dev react app env development ci false npm run build build prod react app env production ci true npm run build build android cd android gradlew bundlerelease test jest passwithnotests cy open node env test cypress open cy run all node env test cypress run config video false cy run smoke node env test cypress run config video false integrationfolder cypress integration smoke serve serve s build p serve test cypress start test serve cy run all dependencies react react native react native vector icons react native web react scripts devdependencies babel core babel plugin proposal class properties babel runtime cypress webpack preprocessor babel jest babel plugin react native web customize cra cypress husky jest react app rewired react test renderer serve start server and test jest preset react native collectcoveragefrom src js jsx mjs testenvironment node testurl modulenamemapper react native react native web husky hooks pre commit npm run lint browserslist production not dead not op mini all development last chrome version last firefox version last safari version config overrides js javascript const override disableeslint babelinclude addbabelplugin addexternalbabelplugins addwebpackplugin require customize cra const webpack require webpack const path require path module exports override disableeslint addbabelplugin react native web addexternalbabelplugins babel plugin proposal class properties babelinclude path resolve src path resolve node modules react native vector icons addwebpackplugin new webpack defineplugin dev process env react app env development process env node env development operating system ubuntu cypress version browser version
| 1
|
86,258
| 8,030,274,333
|
IssuesEvent
|
2018-07-27 18:59:23
|
Microsoft/vscode
|
https://api.github.com/repos/Microsoft/vscode
|
opened
|
Test: Clickable links tsconfig.json for project references
|
testplan-item
|
Test for #54104
- [ ] Mac
- [ ] Linux
- [ ] Windows
Complexity: 1
**Summary**
With TS 3.0, tsconfig.json files may also specify project referecenes.
```js
{
"references": [
{ "path": "./other/project/tsconfig.app.json" }
]
}
```
This references field can either be a path to a json file, or a path to a folder that contains a `tsconfig.json` folder
The project references paths should now be clickable links in VS code.
**Test**
Test that both types of links (links to tsconfigs and to folders containing tsconfigs) are clickable in VS Code and open the correct file when clicked. All paths should be resolved relative to the current tsconfig.json
|
1.0
|
Test: Clickable links tsconfig.json for project references - Test for #54104
- [ ] Mac
- [ ] Linux
- [ ] Windows
Complexity: 1
**Summary**
With TS 3.0, tsconfig.json files may also specify project referecenes.
```js
{
"references": [
{ "path": "./other/project/tsconfig.app.json" }
]
}
```
This references field can either be a path to a json file, or a path to a folder that contains a `tsconfig.json` folder
The project references paths should now be clickable links in VS code.
**Test**
Test that both types of links (links to tsconfigs and to folders containing tsconfigs) are clickable in VS Code and open the correct file when clicked. All paths should be resolved relative to the current tsconfig.json
|
non_process
|
test clickable links tsconfig json for project references test for mac linux windows complexity summary with ts tsconfig json files may also specify project referecenes js references path other project tsconfig app json this references field can either be a path to a json file or a path to a folder that contains a tsconfig json folder the project references paths should now be clickable links in vs code test test that both types of links links to tsconfigs and to folders containing tsconfigs are clickable in vs code and open the correct file when clicked all paths should be resolved relative to the current tsconfig json
| 0
|
16,884
| 22,162,776,328
|
IssuesEvent
|
2022-06-04 19:04:02
|
maticnetwork/miden
|
https://api.github.com/repos/maticnetwork/miden
|
opened
|
Add operation tracking to VmStateIterator
|
processor
|
Currently, the [VmState](https://github.com/maticnetwork/miden/blob/next/processor/src/debug.rs#L7) struct which is returned from the `VmStateIterator` does not contain the operation which was executed to put the VM into this state. We should add another field to this struct so that it looks something like this:
```Rust
#[derive(Clone, Debug, Eq, PartialEq)]
pub struct VmState {
pub op: Operation,
pub clk: usize,
pub fmp: Felt,
pub stack: Vec<Felt>,
pub memory: Vec<(u64, Word)>,
}
```
The `op` field would need to be populated from the information in the [decoder](https://github.com/maticnetwork/miden/blob/next/processor/src/decoder/mod.rs#L209). The decoder doesn't explicitly track the operations yet. It is possible to infer the operations from the trace, but I think a better approach would be to have a vector of operations in the decoder struct. Then, as the VM executes operations, they would be pushed into this vector.
Since tracking operations would result in some overhead, the above should happen only when we are executing programs via [execute_iter()](https://github.com/maticnetwork/miden/blob/next/processor/src/lib.rs#L80) function.
Implementing this functionality well enable counting operations executed by the VM which would be useful for things like #198.
|
1.0
|
Add operation tracking to VmStateIterator - Currently, the [VmState](https://github.com/maticnetwork/miden/blob/next/processor/src/debug.rs#L7) struct which is returned from the `VmStateIterator` does not contain the operation which was executed to put the VM into this state. We should add another field to this struct so that it looks something like this:
```Rust
#[derive(Clone, Debug, Eq, PartialEq)]
pub struct VmState {
pub op: Operation,
pub clk: usize,
pub fmp: Felt,
pub stack: Vec<Felt>,
pub memory: Vec<(u64, Word)>,
}
```
The `op` field would need to be populated from the information in the [decoder](https://github.com/maticnetwork/miden/blob/next/processor/src/decoder/mod.rs#L209). The decoder doesn't explicitly track the operations yet. It is possible to infer the operations from the trace, but I think a better approach would be to have a vector of operations in the decoder struct. Then, as the VM executes operations, they would be pushed into this vector.
Since tracking operations would result in some overhead, the above should happen only when we are executing programs via [execute_iter()](https://github.com/maticnetwork/miden/blob/next/processor/src/lib.rs#L80) function.
Implementing this functionality well enable counting operations executed by the VM which would be useful for things like #198.
|
process
|
add operation tracking to vmstateiterator currently the struct which is returned from the vmstateiterator does not contain the operation which was executed to put the vm into this state we should add another field to this struct so that it looks something like this rust pub struct vmstate pub op operation pub clk usize pub fmp felt pub stack vec pub memory vec the op field would need to be populated from the information in the the decoder doesn t explicitly track the operations yet it is possible to infer the operations from the trace but i think a better approach would be to have a vector of operations in the decoder struct then as the vm executes operations they would be pushed into this vector since tracking operations would result in some overhead the above should happen only when we are executing programs via function implementing this functionality well enable counting operations executed by the vm which would be useful for things like
| 1
|
259,253
| 27,621,778,618
|
IssuesEvent
|
2023-03-10 01:10:59
|
praneethpanasala/linux
|
https://api.github.com/repos/praneethpanasala/linux
|
opened
|
CVE-2023-1074 (Medium) detected in linuxlinux-4.19.6
|
Mend: dependency security vulnerability
|
## CVE-2023-1074 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.6</b></p></summary>
<p>
<p>Apache Software Foundation (ASF)</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/sctp/bind_addr.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/sctp/bind_addr.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This CVE was marked as RESERVED by NVD. This means that the CVE-ID is reserved for future use by the CVE Numbering Authority (CNA) or a security researcher. Still, the CVE details were not included in the CVE entry. This is because the original requester of the CVE ID assignment has not sent an update to MITRE with the information needed to populate the CVE entry.
This page will reflect the classification results once they are available through NVD
<p>Publish Date: 2023-02-28
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-1074>CVE-2023-1074</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-1074">https://www.linuxkernelcves.com/cves/CVE-2023-1074</a></p>
<p>Release Date: 2023-02-28</p>
<p>Fix Resolution: v4.14.305,v4.19.272,v5.4.231,v5.10.166,v5.15.91,v6.1.9,v6.2-rc6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2023-1074 (Medium) detected in linuxlinux-4.19.6 - ## CVE-2023-1074 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.6</b></p></summary>
<p>
<p>Apache Software Foundation (ASF)</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/sctp/bind_addr.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/sctp/bind_addr.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This CVE was marked as RESERVED by NVD. This means that the CVE-ID is reserved for future use by the CVE Numbering Authority (CNA) or a security researcher. Still, the CVE details were not included in the CVE entry. This is because the original requester of the CVE ID assignment has not sent an update to MITRE with the information needed to populate the CVE entry.
This page will reflect the classification results once they are available through NVD
<p>Publish Date: 2023-02-28
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-1074>CVE-2023-1074</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-1074">https://www.linuxkernelcves.com/cves/CVE-2023-1074</a></p>
<p>Release Date: 2023-02-28</p>
<p>Fix Resolution: v4.14.305,v4.19.272,v5.4.231,v5.10.166,v5.15.91,v6.1.9,v6.2-rc6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_process
|
cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux apache software foundation asf library home page a href found in base branch master vulnerable source files net sctp bind addr c net sctp bind addr c vulnerability details this cve was marked as reserved by nvd this means that the cve id is reserved for future use by the cve numbering authority cna or a security researcher still the cve details were not included in the cve entry this is because the original requester of the cve id assignment has not sent an update to mitre with the information needed to populate the cve entry this page will reflect the classification results once they are available through nvd publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
| 0
|
199,916
| 6,996,123,204
|
IssuesEvent
|
2017-12-15 22:30:22
|
freechipsproject/firrtl
|
https://api.github.com/repos/freechipsproject/firrtl
|
closed
|
Emit Verilog Top-Down
|
enhancement low priority
|
So recently, when I was debugging some emitted Verilog, I kept getting tripped up by the fact that logical dependencies kind of went from bottom to top. It doesn't matter as far as synthesizability is concerned, because the signals are all pre-declared, but for some debugging sanity, it'd be super nice if Verilog was emitted top to bottom so the logical flow made more sense.
|
1.0
|
Emit Verilog Top-Down - So recently, when I was debugging some emitted Verilog, I kept getting tripped up by the fact that logical dependencies kind of went from bottom to top. It doesn't matter as far as synthesizability is concerned, because the signals are all pre-declared, but for some debugging sanity, it'd be super nice if Verilog was emitted top to bottom so the logical flow made more sense.
|
non_process
|
emit verilog top down so recently when i was debugging some emitted verilog i kept getting tripped up by the fact that logical dependencies kind of went from bottom to top it doesn t matter as far as synthesizability is concerned because the signals are all pre declared but for some debugging sanity it d be super nice if verilog was emitted top to bottom so the logical flow made more sense
| 0
|
37,749
| 5,142,132,898
|
IssuesEvent
|
2017-01-12 12:15:02
|
kubernetes/kubernetes
|
https://api.github.com/repos/kubernetes/kubernetes
|
opened
|
ci-kubernetes-e2e-gci-gke-subnet: broken test run
|
kind/flake priority/P2 team/test-infra
|
https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-e2e-gci-gke-subnet/191/
Multiple broken tests:
Failed: [k8s.io] DNS should provide DNS for the cluster [Conformance] {Kubernetes e2e suite}
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/dns.go:352
Expected error:
<*errors.errorString | 0xc42038cc30>: {
s: "timed out waiting for the condition",
}
timed out waiting for the condition
not to have occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/dns.go:236
```
Issues about this test specifically: #26194 #26338 #30345 #34571
Failed: [k8s.io] DNS config map should be able to change configuration {Kubernetes e2e suite}
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/dns_configmap.go:66
Expected error:
<*errors.errorString | 0xc42038cc30>: {
s: "timed out waiting for the condition",
}
timed out waiting for the condition
not to have occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/dns_configmap.go:283
```
Issues about this test specifically: #37144
Failed: [k8s.io] DNS should provide DNS for ExternalName services {Kubernetes e2e suite}
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/dns.go:501
Expected error:
<*errors.errorString | 0xc42038cc30>: {
s: "timed out waiting for the condition",
}
timed out waiting for the condition
not to have occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/dns.go:265
```
Issues about this test specifically: #32584
Failed: Test {e2e.go}
```
exit status 1
```
Issues about this test specifically: #33361 #38663 #39788
Previous issues for this suite: #37341 #38356 #38582 #39186
|
1.0
|
ci-kubernetes-e2e-gci-gke-subnet: broken test run - https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-e2e-gci-gke-subnet/191/
Multiple broken tests:
Failed: [k8s.io] DNS should provide DNS for the cluster [Conformance] {Kubernetes e2e suite}
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/dns.go:352
Expected error:
<*errors.errorString | 0xc42038cc30>: {
s: "timed out waiting for the condition",
}
timed out waiting for the condition
not to have occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/dns.go:236
```
Issues about this test specifically: #26194 #26338 #30345 #34571
Failed: [k8s.io] DNS config map should be able to change configuration {Kubernetes e2e suite}
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/dns_configmap.go:66
Expected error:
<*errors.errorString | 0xc42038cc30>: {
s: "timed out waiting for the condition",
}
timed out waiting for the condition
not to have occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/dns_configmap.go:283
```
Issues about this test specifically: #37144
Failed: [k8s.io] DNS should provide DNS for ExternalName services {Kubernetes e2e suite}
```
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/dns.go:501
Expected error:
<*errors.errorString | 0xc42038cc30>: {
s: "timed out waiting for the condition",
}
timed out waiting for the condition
not to have occurred
/go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/dns.go:265
```
Issues about this test specifically: #32584
Failed: Test {e2e.go}
```
exit status 1
```
Issues about this test specifically: #33361 #38663 #39788
Previous issues for this suite: #37341 #38356 #38582 #39186
|
non_process
|
ci kubernetes gci gke subnet broken test run multiple broken tests failed dns should provide dns for the cluster kubernetes suite go src io kubernetes output dockerized go src io kubernetes test dns go expected error s timed out waiting for the condition timed out waiting for the condition not to have occurred go src io kubernetes output dockerized go src io kubernetes test dns go issues about this test specifically failed dns config map should be able to change configuration kubernetes suite go src io kubernetes output dockerized go src io kubernetes test dns configmap go expected error s timed out waiting for the condition timed out waiting for the condition not to have occurred go src io kubernetes output dockerized go src io kubernetes test dns configmap go issues about this test specifically failed dns should provide dns for externalname services kubernetes suite go src io kubernetes output dockerized go src io kubernetes test dns go expected error s timed out waiting for the condition timed out waiting for the condition not to have occurred go src io kubernetes output dockerized go src io kubernetes test dns go issues about this test specifically failed test go exit status issues about this test specifically previous issues for this suite
| 0
|
4,623
| 7,468,690,066
|
IssuesEvent
|
2018-04-02 19:54:44
|
w3c/w3process
|
https://api.github.com/repos/w3c/w3process
|
closed
|
Process2014 introduced an AC ballot for CR transitions and now we have a ballot open for a year
|
Process2018Candidate
|
Transferred from https://www.w3.org/community/w3process/track/issues/182
State: Raised
|
1.0
|
Process2014 introduced an AC ballot for CR transitions and now we have a ballot open for a year - Transferred from https://www.w3.org/community/w3process/track/issues/182
State: Raised
|
process
|
introduced an ac ballot for cr transitions and now we have a ballot open for a year transferred from state raised
| 1
|
12,359
| 14,888,287,582
|
IssuesEvent
|
2021-01-20 19:37:06
|
CodeForPittsburgh/food-access-map-data
|
https://api.github.com/repos/CodeForPittsburgh/food-access-map-data
|
closed
|
process new farmers market data
|
data processing
|
the new farmers market datasource needs to be wrangled into our schema.
|
1.0
|
process new farmers market data - the new farmers market datasource needs to be wrangled into our schema.
|
process
|
process new farmers market data the new farmers market datasource needs to be wrangled into our schema
| 1
|
61,118
| 6,725,781,409
|
IssuesEvent
|
2017-10-17 07:25:35
|
openshift/origin
|
https://api.github.com/repos/openshift/origin
|
closed
|
flake: templateservicebroker security test should pass security tests
|
area/security kind/test-flake priority/P0
|
https://openshift-gce-devel.appspot.com/build/origin-ci-test/pr-logs/pull/16851/test_pull_request_origin_extended_conformance_gce/9755/
```
[Conformance][templates] templateservicebroker security test should pass security tests [Suite:openshift/conformance/parallel] 1m43s
go run hack/e2e.go -v -test --test_args='--ginkgo.focus=\s\[Conformance\]\[templates\]\stemplateservicebroker\ssecurity\stest\sshould\spass\ssecurity\stests\s\[Suite\:openshift\/conformance\/parallel\]$'
/tmp/openshift/build-rpm-release/tito/rpmbuild-origin97017y/BUILD/origin-3.7.0/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateservicebroker_security.go:257
Expected error:
<*client.ServerError | 0xc421d8b3c0>: {
StatusCode: 403,
Description: "secrets \"d2a2b8b4-b577-42c8-92b7-0f5b85d03592\" is forbidden: cannot set blockOwnerDeletion if an ownerReference refers to a resource you can't set finalizers on: User \"system:serviceaccount:extended-test-openshift-template-service-broker-j0pr1-hp4dk:apiserver\" cannot update brokertemplateinstances/finalizers.template.openshift.io in project \"extended-test-templates-46xbx-whdrr\", <nil>",
}
Forbidden: secrets "d2a2b8b4-b577-42c8-92b7-0f5b85d03592" is forbidden: cannot set blockOwnerDeletion if an ownerReference refers to a resource you can't set finalizers on: User "system:serviceaccount:extended-test-openshift-template-service-broker-j0pr1-hp4dk:apiserver" cannot update brokertemplateinstances/finalizers.template.openshift.io in project "extended-test-templates-46xbx-whdrr", <nil>
not to have occurred
/tmp/openshift/build-rpm-release/tito/rpmbuild-origin97017y/BUILD/origin-3.7.0/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateservicebroker_security.go:162
```
Looks like permission cache not being filled yet?
@openshift/sig-security @openshift/sig-master
|
1.0
|
flake: templateservicebroker security test should pass security tests - https://openshift-gce-devel.appspot.com/build/origin-ci-test/pr-logs/pull/16851/test_pull_request_origin_extended_conformance_gce/9755/
```
[Conformance][templates] templateservicebroker security test should pass security tests [Suite:openshift/conformance/parallel] 1m43s
go run hack/e2e.go -v -test --test_args='--ginkgo.focus=\s\[Conformance\]\[templates\]\stemplateservicebroker\ssecurity\stest\sshould\spass\ssecurity\stests\s\[Suite\:openshift\/conformance\/parallel\]$'
/tmp/openshift/build-rpm-release/tito/rpmbuild-origin97017y/BUILD/origin-3.7.0/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateservicebroker_security.go:257
Expected error:
<*client.ServerError | 0xc421d8b3c0>: {
StatusCode: 403,
Description: "secrets \"d2a2b8b4-b577-42c8-92b7-0f5b85d03592\" is forbidden: cannot set blockOwnerDeletion if an ownerReference refers to a resource you can't set finalizers on: User \"system:serviceaccount:extended-test-openshift-template-service-broker-j0pr1-hp4dk:apiserver\" cannot update brokertemplateinstances/finalizers.template.openshift.io in project \"extended-test-templates-46xbx-whdrr\", <nil>",
}
Forbidden: secrets "d2a2b8b4-b577-42c8-92b7-0f5b85d03592" is forbidden: cannot set blockOwnerDeletion if an ownerReference refers to a resource you can't set finalizers on: User "system:serviceaccount:extended-test-openshift-template-service-broker-j0pr1-hp4dk:apiserver" cannot update brokertemplateinstances/finalizers.template.openshift.io in project "extended-test-templates-46xbx-whdrr", <nil>
not to have occurred
/tmp/openshift/build-rpm-release/tito/rpmbuild-origin97017y/BUILD/origin-3.7.0/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateservicebroker_security.go:162
```
Looks like permission cache not being filled yet?
@openshift/sig-security @openshift/sig-master
|
non_process
|
flake templateservicebroker security test should pass security tests templateservicebroker security test should pass security tests go run hack go v test test args ginkgo focus s stemplateservicebroker ssecurity stest sshould spass ssecurity stests s tmp openshift build rpm release tito rpmbuild build origin output local go src github com openshift origin test extended templates templateservicebroker security go expected error statuscode description secrets is forbidden cannot set blockownerdeletion if an ownerreference refers to a resource you can t set finalizers on user system serviceaccount extended test openshift template service broker apiserver cannot update brokertemplateinstances finalizers template openshift io in project extended test templates whdrr forbidden secrets is forbidden cannot set blockownerdeletion if an ownerreference refers to a resource you can t set finalizers on user system serviceaccount extended test openshift template service broker apiserver cannot update brokertemplateinstances finalizers template openshift io in project extended test templates whdrr not to have occurred tmp openshift build rpm release tito rpmbuild build origin output local go src github com openshift origin test extended templates templateservicebroker security go looks like permission cache not being filled yet openshift sig security openshift sig master
| 0
|
17,622
| 23,442,524,939
|
IssuesEvent
|
2022-08-15 16:13:24
|
pytorch/pytorch
|
https://api.github.com/repos/pytorch/pytorch
|
closed
|
Can't pickle local object 'CDLL.__init__.<locals>._FuncPtr'
|
module: multiprocessing triaged
|
I am using `torch.multiprocessing.Process` in a reinforcement learning project. Here's bit of my codes:
```python
class Agent(mp.Process):
def __init__():
super()...
self.dll = CDLL('load a dll here which work fine without mp.Process')
self.env = some_env
def run(self):
action = ...
self.env.step(action, self.dll)
...
if __name__ == "__main__":
mp.set_start_method("forkserver")
workers = [Agent(..) for i in range(n)]
[w.start() for w in workers]
[w.join() for w in workers]
```
Here's the errors I got:
```
Traceback (most recent call last):
File "gym_test.py", line 173, in <module>
[w.start() for w in workers]
File "gym_test.py", line 173, in <listcomp>
[w.start() for w in workers]
File "/usr/lib/python3.7/multiprocessing/process.py", line 112, in start
self._popen = self._Popen(self)
File "/usr/lib/python3.7/multiprocessing/context.py", line 223, in _Popen
return _default_context.get_context().Process._Popen(process_obj)
File "/usr/lib/python3.7/multiprocessing/context.py", line 291, in _Popen
return Popen(process_obj)
File "/usr/lib/python3.7/multiprocessing/popen_forkserver.py", line 35, in __init__
super().__init__(process_obj)
File "/usr/lib/python3.7/multiprocessing/popen_fork.py", line 20, in __init__
self._launch(process_obj)
File "/usr/lib/python3.7/multiprocessing/popen_forkserver.py", line 47, in _launch
reduction.dump(process_obj, buf)
File "/usr/lib/python3.7/multiprocessing/reduction.py", line 60, in dump
ForkingPickler(file, protocol).dump(obj)
AttributeError: Can't pickle local object 'CDLL.__init__.<locals>._FuncPtr'
```
I wonder how to load cdll properly in multiprocessing, any solutions?
cc @VitalyFedyunin
|
1.0
|
Can't pickle local object 'CDLL.__init__.<locals>._FuncPtr' - I am using `torch.multiprocessing.Process` in a reinforcement learning project. Here's bit of my codes:
```python
class Agent(mp.Process):
def __init__():
super()...
self.dll = CDLL('load a dll here which work fine without mp.Process')
self.env = some_env
def run(self):
action = ...
self.env.step(action, self.dll)
...
if __name__ == "__main__":
mp.set_start_method("forkserver")
workers = [Agent(..) for i in range(n)]
[w.start() for w in workers]
[w.join() for w in workers]
```
Here's the errors I got:
```
Traceback (most recent call last):
File "gym_test.py", line 173, in <module>
[w.start() for w in workers]
File "gym_test.py", line 173, in <listcomp>
[w.start() for w in workers]
File "/usr/lib/python3.7/multiprocessing/process.py", line 112, in start
self._popen = self._Popen(self)
File "/usr/lib/python3.7/multiprocessing/context.py", line 223, in _Popen
return _default_context.get_context().Process._Popen(process_obj)
File "/usr/lib/python3.7/multiprocessing/context.py", line 291, in _Popen
return Popen(process_obj)
File "/usr/lib/python3.7/multiprocessing/popen_forkserver.py", line 35, in __init__
super().__init__(process_obj)
File "/usr/lib/python3.7/multiprocessing/popen_fork.py", line 20, in __init__
self._launch(process_obj)
File "/usr/lib/python3.7/multiprocessing/popen_forkserver.py", line 47, in _launch
reduction.dump(process_obj, buf)
File "/usr/lib/python3.7/multiprocessing/reduction.py", line 60, in dump
ForkingPickler(file, protocol).dump(obj)
AttributeError: Can't pickle local object 'CDLL.__init__.<locals>._FuncPtr'
```
I wonder how to load cdll properly in multiprocessing, any solutions?
cc @VitalyFedyunin
|
process
|
can t pickle local object cdll init funcptr i am using torch multiprocessing process in a reinforcement learning project here s bit of my codes python class agent mp process def init super self dll cdll load a dll here which work fine without mp process self env some env def run self action self env step action self dll if name main mp set start method forkserver workers here s the errors i got traceback most recent call last file gym test py line in file gym test py line in file usr lib multiprocessing process py line in start self popen self popen self file usr lib multiprocessing context py line in popen return default context get context process popen process obj file usr lib multiprocessing context py line in popen return popen process obj file usr lib multiprocessing popen forkserver py line in init super init process obj file usr lib multiprocessing popen fork py line in init self launch process obj file usr lib multiprocessing popen forkserver py line in launch reduction dump process obj buf file usr lib multiprocessing reduction py line in dump forkingpickler file protocol dump obj attributeerror can t pickle local object cdll init funcptr i wonder how to load cdll properly in multiprocessing any solutions cc vitalyfedyunin
| 1
|
50,373
| 6,367,231,791
|
IssuesEvent
|
2017-08-01 05:20:58
|
openMF/community-app
|
https://api.github.com/repos/openMF/community-app
|
closed
|
Charge Fee Frequency is not displaying properly.
|
bug design gsoc p1
|
1. Go to Admin>Products>Charges and Create Charge
2. Select Loan for "Charge Applies To"
3. Furnish the inputs and select 'Charge Time Type' as Overdue Fees
4. Enable 'Add Frequency' and click on drop-down menu of Charge Frequency.
>Check the following Screenshot:

Expected: It should just display Days, Weeks, Months and Years.
|
1.0
|
Charge Fee Frequency is not displaying properly. - 1. Go to Admin>Products>Charges and Create Charge
2. Select Loan for "Charge Applies To"
3. Furnish the inputs and select 'Charge Time Type' as Overdue Fees
4. Enable 'Add Frequency' and click on drop-down menu of Charge Frequency.
>Check the following Screenshot:

Expected: It should just display Days, Weeks, Months and Years.
|
non_process
|
charge fee frequency is not displaying properly go to admin products charges and create charge select loan for charge applies to furnish the inputs and select charge time type as overdue fees enable add frequency and click on drop down menu of charge frequency check the following screenshot expected it should just display days weeks months and years
| 0
|
795,891
| 28,091,058,910
|
IssuesEvent
|
2023-03-30 13:05:42
|
webcompat/web-bugs
|
https://api.github.com/repos/webcompat/web-bugs
|
closed
|
twitch.tv - site is not usable
|
browser-firefox priority-important os-linux engine-gecko
|
<!-- @browser: Firefox 111.0 -->
<!-- @ua_header: Mozilla/5.0 (X11; Linux x86_64; rv:109.0) Gecko/20100101 Firefox/111.0 -->
<!-- @reported_with: unknown -->
**URL**: https://twitch.tv
**Browser / Version**: Firefox 111.0
**Operating System**: Linux
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Browser unsupported
**Steps to Reproduce**:
Login page says browser is unsupported
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2023/3/22c79761-d36e-4de1-911a-f6f4bba1b27e.jpg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
1.0
|
twitch.tv - site is not usable - <!-- @browser: Firefox 111.0 -->
<!-- @ua_header: Mozilla/5.0 (X11; Linux x86_64; rv:109.0) Gecko/20100101 Firefox/111.0 -->
<!-- @reported_with: unknown -->
**URL**: https://twitch.tv
**Browser / Version**: Firefox 111.0
**Operating System**: Linux
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Browser unsupported
**Steps to Reproduce**:
Login page says browser is unsupported
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2023/3/22c79761-d36e-4de1-911a-f6f4bba1b27e.jpg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_
|
non_process
|
twitch tv site is not usable url browser version firefox operating system linux tested another browser yes chrome problem type site is not usable description browser unsupported steps to reproduce login page says browser is unsupported view the screenshot img alt screenshot src browser configuration none from with ❤️
| 0
|
6,622
| 9,725,389,829
|
IssuesEvent
|
2019-05-30 08:34:12
|
qgis/QGIS
|
https://api.github.com/repos/qgis/QGIS
|
closed
|
Scan data for attributes to use in a Processing model
|
Feature Request Processing
|
Author Name: **Magnus Nilsson** (Magnus Nilsson)
Original Redmine Issue: [20986](https://issues.qgis.org/issues/20986)
Redmine category:processing/modeller
---
I am not too fond of the idea of manually having to enter which attributes to use in a Processing model. For example, in the tool for creating points from a table, I manually have to enter (write) which fields to use for x/y coordinates. Wouldn´t it be better to allow a model to scan the data for attributes and let the user select relevant attributes from a list?
Related: When adding a vector layer to a model, why not choose the data source directly? That way, the model knows the available attributes. Having to define this each time I run a model is time consuming.
|
1.0
|
Scan data for attributes to use in a Processing model - Author Name: **Magnus Nilsson** (Magnus Nilsson)
Original Redmine Issue: [20986](https://issues.qgis.org/issues/20986)
Redmine category:processing/modeller
---
I am not too fond of the idea of manually having to enter which attributes to use in a Processing model. For example, in the tool for creating points from a table, I manually have to enter (write) which fields to use for x/y coordinates. Wouldn´t it be better to allow a model to scan the data for attributes and let the user select relevant attributes from a list?
Related: When adding a vector layer to a model, why not choose the data source directly? That way, the model knows the available attributes. Having to define this each time I run a model is time consuming.
|
process
|
scan data for attributes to use in a processing model author name magnus nilsson magnus nilsson original redmine issue redmine category processing modeller i am not too fond of the idea of manually having to enter which attributes to use in a processing model for example in the tool for creating points from a table i manually have to enter write which fields to use for x y coordinates wouldn´t it be better to allow a model to scan the data for attributes and let the user select relevant attributes from a list related when adding a vector layer to a model why not choose the data source directly that way the model knows the available attributes having to define this each time i run a model is time consuming
| 1
|
131,113
| 27,824,457,756
|
IssuesEvent
|
2023-03-19 15:58:25
|
pinterest/ktlint
|
https://api.github.com/repos/pinterest/ktlint
|
closed
|
Indentation change since 0.38.1
|
indentation-rule conflict-with-default-intellij-formatting ktlint-official-codestyle
|
## Expected Behavior
Upgrading from 0.38.1 to latest I see a change in indentation with parameter names in a call.
For example this was allowed in 0.38.1:
someFunction(
parameterName =
someValue
.someProperty
.someCall()
)
## Observed Behavior
With 0.41.1 it now demands that it be indented like this:
someFunction(
parameterName =
someValue
.someProperty
.someCall()
)
Which is not very readable.
I can make it go away by moving someValue to same line as parameter name:
someFunction(
parameterName = someValue
.someProperty
.someCall()
)
But sometimes there is a longer expression than this example and it is more readable to move it to its own line.
|
1.0
|
Indentation change since 0.38.1 - ## Expected Behavior
Upgrading from 0.38.1 to latest I see a change in indentation with parameter names in a call.
For example this was allowed in 0.38.1:
someFunction(
parameterName =
someValue
.someProperty
.someCall()
)
## Observed Behavior
With 0.41.1 it now demands that it be indented like this:
someFunction(
parameterName =
someValue
.someProperty
.someCall()
)
Which is not very readable.
I can make it go away by moving someValue to same line as parameter name:
someFunction(
parameterName = someValue
.someProperty
.someCall()
)
But sometimes there is a longer expression than this example and it is more readable to move it to its own line.
|
non_process
|
indentation change since expected behavior upgrading from to latest i see a change in indentation with parameter names in a call for example this was allowed in somefunction parametername somevalue someproperty somecall observed behavior with it now demands that it be indented like this somefunction parametername somevalue someproperty somecall which is not very readable i can make it go away by moving somevalue to same line as parameter name somefunction parametername somevalue someproperty somecall but sometimes there is a longer expression than this example and it is more readable to move it to its own line
| 0
|
3,331
| 6,447,850,646
|
IssuesEvent
|
2017-08-14 09:22:41
|
nodejs/node
|
https://api.github.com/repos/nodejs/node
|
closed
|
maxBuffer default too small
|
child_process
|
* **Version**: v6.2.1
* **Platform**: Darwin me.local 15.5.0 Darwin Kernel Version 15.5.0: Tue Apr 19 18:36:36 PDT 2016; root:xnu-3248.50.21~8/RELEASE_X86_64 x86_64
* **Subsystem**: child_process
Currently `maxBuffer` for `child_process.exec` is set to 200*1024 bytes, or ~204.8KB. I ran into an issue where my child process was being terminated and tracking it down was quite tough. It ended up being that it was producing enough output that it exceeded `maxBuffer`.
I think the buffer size is too small and this behavior (terminating a child) is drastic enough that it should only be done in the case where a child is producing a much larger amount of output.
I'm not sure what's sane here, perhaps 5MB+?
|
1.0
|
maxBuffer default too small - * **Version**: v6.2.1
* **Platform**: Darwin me.local 15.5.0 Darwin Kernel Version 15.5.0: Tue Apr 19 18:36:36 PDT 2016; root:xnu-3248.50.21~8/RELEASE_X86_64 x86_64
* **Subsystem**: child_process
Currently `maxBuffer` for `child_process.exec` is set to 200*1024 bytes, or ~204.8KB. I ran into an issue where my child process was being terminated and tracking it down was quite tough. It ended up being that it was producing enough output that it exceeded `maxBuffer`.
I think the buffer size is too small and this behavior (terminating a child) is drastic enough that it should only be done in the case where a child is producing a much larger amount of output.
I'm not sure what's sane here, perhaps 5MB+?
|
process
|
maxbuffer default too small version platform darwin me local darwin kernel version tue apr pdt root xnu release subsystem child process currently maxbuffer for child process exec is set to bytes or i ran into an issue where my child process was being terminated and tracking it down was quite tough it ended up being that it was producing enough output that it exceeded maxbuffer i think the buffer size is too small and this behavior terminating a child is drastic enough that it should only be done in the case where a child is producing a much larger amount of output i m not sure what s sane here perhaps
| 1
|
728,712
| 25,089,164,885
|
IssuesEvent
|
2022-11-08 03:53:08
|
ballerina-platform/ballerina-dev-website
|
https://api.github.com/repos/ballerina-platform/ballerina-dev-website
|
closed
|
Improve styles in newsletter
|
Priority/Highest Type/Improvement
|
## Description
> $subject
## Related website/documentation area
> Add/Uncomment the relevant area label out of the following.
<!--Area/BBEs-->
<!--Area/HomePageSamples-->
<!--Area/LearnPages-->
<!--Area/CommonPages-->
<!--Area/Backend-->
<!--Area/UIUX-->
<!--Area/Workflows-->
<!--Area/Blog-->
## Describe the problem(s)
> A detailed description of the purpose of this improvement.
## Describe your solution(s)
> A detailed description of how this improvement will solve the problem described above.
## Related issue(s) (optional)
> Any related issues such as sub tasks and issues reported in other repositories (e.g., component repositories), similar problems, etc.
## Suggested label(s) (optional)
> Optional comma-separated list of suggested labels. Non committers can’t assign labels to issues, and thereby, this will help issue creators who are not a committer to suggest possible labels.
## Suggested assignee(s) (optional)
> Optional comma-separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, and thereby, this will help issue creators who are not a committer to suggest possible assignees.
|
1.0
|
Improve styles in newsletter - ## Description
> $subject
## Related website/documentation area
> Add/Uncomment the relevant area label out of the following.
<!--Area/BBEs-->
<!--Area/HomePageSamples-->
<!--Area/LearnPages-->
<!--Area/CommonPages-->
<!--Area/Backend-->
<!--Area/UIUX-->
<!--Area/Workflows-->
<!--Area/Blog-->
## Describe the problem(s)
> A detailed description of the purpose of this improvement.
## Describe your solution(s)
> A detailed description of how this improvement will solve the problem described above.
## Related issue(s) (optional)
> Any related issues such as sub tasks and issues reported in other repositories (e.g., component repositories), similar problems, etc.
## Suggested label(s) (optional)
> Optional comma-separated list of suggested labels. Non committers can’t assign labels to issues, and thereby, this will help issue creators who are not a committer to suggest possible labels.
## Suggested assignee(s) (optional)
> Optional comma-separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, and thereby, this will help issue creators who are not a committer to suggest possible assignees.
|
non_process
|
improve styles in newsletter description subject related website documentation area add uncomment the relevant area label out of the following describe the problem s a detailed description of the purpose of this improvement describe your solution s a detailed description of how this improvement will solve the problem described above related issue s optional any related issues such as sub tasks and issues reported in other repositories e g component repositories similar problems etc suggested label s optional optional comma separated list of suggested labels non committers can’t assign labels to issues and thereby this will help issue creators who are not a committer to suggest possible labels suggested assignee s optional optional comma separated list of suggested team members who should attend the issue non committers can’t assign issues to assignees and thereby this will help issue creators who are not a committer to suggest possible assignees
| 0
|
448,689
| 31,809,004,675
|
IssuesEvent
|
2023-09-13 15:34:20
|
vijayk3327/Visualforce
|
https://api.github.com/repos/vijayk3327/Visualforce
|
opened
|
How to Get Billing Address of Account Object from Quote based on record Id using Apex Class Method in Visualforce Page Salesforce
|
documentation question
|
In this post we are going to learn about How to Get Billing Address of Account Object from Quote based on record Id using Apex Class Method in Visualforce Page Salesforce.
Address is a “compound field” – in simple terms, this means the field groups together multiple individual elements into a single compound.
Custom Address Fields:- Users can populate a custom address fields manually or they can use the Google lookup to search for an address. Admins and APIs can access each address stored in a custom address field as a structured compound data type as well as individual address components.
**[👉 Get source code live demo link:-](https://www.w3web.net/get-account-billing-address-from-quote/)**
<img src="https://www.w3web.net/wp-content/uploads/2023/05/accAddressFromQuote-min.gif"/>
`Create Visualforce Page
Step 1:- Create Visualforce Page : accAddressFromQuoteVf.vfp`
`<apex:page standardController="Quote" extensions="accAddressFromQuoteVfCtrl">
<div class="slds slds-p-around_small" style="padding:10px;">
<table width="50%" border="1" cellspacing="0" cellpadding="5" bordercolor="#ccc" class="slds-table slds-table_bordered slds-table_col-bordered" style="border-collapse:collapse;">
<tr style="background:#ddd;">
<th>Quote Name</th>
<th>Billing Street</th>
<th>Billing City</th>
<th>Billing Country</th>
<th>Billing Postal Code</th>
</tr>
<apex:repeat value="{!quoteObj}" var="qotItem">
<tr>
<td><p><apex:outputText value="{!qotItem.Name}"/></p></td>
<td><p><apex:outputText value="{!qotItem.Account.BillingStreet}"/></p></td>
<td><p><apex:outputText value="{!qotItem.Account.BillingCity}"/></p></td>
<td><p><apex:outputText value="{!qotItem.Account.BillingCountry}"/></p></td>
<td><p><apex:outputText value="{!qotItem.Account.BillingPostalCode}"/></p></td>
</tr>
</apex:repeat>
</table>
</div>
</apex:page>`
**Create Apex Class Extension Controller in Visualforce
Step 2:- Create Apex Class : accAddressFromQuoteVfCtrl.apxc**
`public class accAddressFromQuoteVfCtrl {
public String MstrId{GET;SET;}
public Quote quoteObj{GET;SET;}
public accAddressFromQuoteVfCtrl(ApexPages.StandardController Controller){
MstrId = ApexPages.currentPage().getParameters().get('id');
quoteObj = [SELECT Id, Name, AccountId, Account.Name, Account.BillingStreet, Account.BillingCity, Account.BillingCountry, Account.BillingPostalCode FROM Quote WHERE Id=:MstrId ];
}
}`
**[👉 Get source code live demo link:-](https://www.w3web.net/get-account-billing-address-from-quote/)**
|
1.0
|
How to Get Billing Address of Account Object from Quote based on record Id using Apex Class Method in Visualforce Page Salesforce - In this post we are going to learn about How to Get Billing Address of Account Object from Quote based on record Id using Apex Class Method in Visualforce Page Salesforce.
Address is a “compound field” – in simple terms, this means the field groups together multiple individual elements into a single compound.
Custom Address Fields:- Users can populate a custom address fields manually or they can use the Google lookup to search for an address. Admins and APIs can access each address stored in a custom address field as a structured compound data type as well as individual address components.
**[👉 Get source code live demo link:-](https://www.w3web.net/get-account-billing-address-from-quote/)**
<img src="https://www.w3web.net/wp-content/uploads/2023/05/accAddressFromQuote-min.gif"/>
`Create Visualforce Page
Step 1:- Create Visualforce Page : accAddressFromQuoteVf.vfp`
`<apex:page standardController="Quote" extensions="accAddressFromQuoteVfCtrl">
<div class="slds slds-p-around_small" style="padding:10px;">
<table width="50%" border="1" cellspacing="0" cellpadding="5" bordercolor="#ccc" class="slds-table slds-table_bordered slds-table_col-bordered" style="border-collapse:collapse;">
<tr style="background:#ddd;">
<th>Quote Name</th>
<th>Billing Street</th>
<th>Billing City</th>
<th>Billing Country</th>
<th>Billing Postal Code</th>
</tr>
<apex:repeat value="{!quoteObj}" var="qotItem">
<tr>
<td><p><apex:outputText value="{!qotItem.Name}"/></p></td>
<td><p><apex:outputText value="{!qotItem.Account.BillingStreet}"/></p></td>
<td><p><apex:outputText value="{!qotItem.Account.BillingCity}"/></p></td>
<td><p><apex:outputText value="{!qotItem.Account.BillingCountry}"/></p></td>
<td><p><apex:outputText value="{!qotItem.Account.BillingPostalCode}"/></p></td>
</tr>
</apex:repeat>
</table>
</div>
</apex:page>`
**Create Apex Class Extension Controller in Visualforce
Step 2:- Create Apex Class : accAddressFromQuoteVfCtrl.apxc**
`public class accAddressFromQuoteVfCtrl {
public String MstrId{GET;SET;}
public Quote quoteObj{GET;SET;}
public accAddressFromQuoteVfCtrl(ApexPages.StandardController Controller){
MstrId = ApexPages.currentPage().getParameters().get('id');
quoteObj = [SELECT Id, Name, AccountId, Account.Name, Account.BillingStreet, Account.BillingCity, Account.BillingCountry, Account.BillingPostalCode FROM Quote WHERE Id=:MstrId ];
}
}`
**[👉 Get source code live demo link:-](https://www.w3web.net/get-account-billing-address-from-quote/)**
|
non_process
|
how to get billing address of account object from quote based on record id using apex class method in visualforce page salesforce in this post we are going to learn about how to get billing address of account object from quote based on record id using apex class method in visualforce page salesforce address is a “compound field” – in simple terms this means the field groups together multiple individual elements into a single compound custom address fields users can populate a custom address fields manually or they can use the google lookup to search for an address admins and apis can access each address stored in a custom address field as a structured compound data type as well as individual address components img src create visualforce page step create visualforce page accaddressfromquotevf vfp quote name billing street billing city billing country billing postal code create apex class extension controller in visualforce step create apex class accaddressfromquotevfctrl apxc public class accaddressfromquotevfctrl public string mstrid get set public quote quoteobj get set public accaddressfromquotevfctrl apexpages standardcontroller controller mstrid apexpages currentpage getparameters get id quoteobj
| 0
|
2,314
| 5,135,355,155
|
IssuesEvent
|
2017-01-11 12:03:29
|
jlm2017/jlm-video-subtitles
|
https://api.github.com/repos/jlm2017/jlm-video-subtitles
|
opened
|
[Subtitles] [FR] MÉLENCHON - Déboulé à Tourcoing
|
Language: French Process: [0] Awaiting subtitles
|
# **Video title**
MÉLENCHON - Déboulé à Tourcoing
# **Url**
https://www.youtube.com/watch?v=NNe0bXIJ9cQ&t=593s
# **Youtube subtitles language**
Français
# **Duration**
1:41:42
# **Subtitles URL**
https://www.youtube.com/timedtext_editor?ref=player&action_mde_edit_form=1&lang=fr&tab=captions&ui=hd&v=NNe0bXIJ9cQ&bl=vmp
|
1.0
|
[Subtitles] [FR] MÉLENCHON - Déboulé à Tourcoing - # **Video title**
MÉLENCHON - Déboulé à Tourcoing
# **Url**
https://www.youtube.com/watch?v=NNe0bXIJ9cQ&t=593s
# **Youtube subtitles language**
Français
# **Duration**
1:41:42
# **Subtitles URL**
https://www.youtube.com/timedtext_editor?ref=player&action_mde_edit_form=1&lang=fr&tab=captions&ui=hd&v=NNe0bXIJ9cQ&bl=vmp
|
process
|
mélenchon déboulé à tourcoing video title mélenchon déboulé à tourcoing url youtube subtitles language français duration subtitles url
| 1
|
350,647
| 31,931,880,542
|
IssuesEvent
|
2023-09-19 08:00:07
|
unifyai/ivy
|
https://api.github.com/repos/unifyai/ivy
|
reopened
|
Fix linalg.test_tensorflow_inv
|
TensorFlow Frontend Sub Task Failing Test
|
| | |
|---|---|
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/6005627299/job/16288695179"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/6005627299/job/16288695179"><img src=https://img.shields.io/badge/-failure-red></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/6005627299/job/16288695179"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/6005627299/job/16288695179"><img src=https://img.shields.io/badge/-failure-red></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/6005627299/job/16288695179"><img src=https://img.shields.io/badge/-failure-red></a>
|
1.0
|
Fix linalg.test_tensorflow_inv - | | |
|---|---|
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/6005627299/job/16288695179"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/6005627299/job/16288695179"><img src=https://img.shields.io/badge/-failure-red></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/6005627299/job/16288695179"><img src=https://img.shields.io/badge/-failure-red></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/6005627299/job/16288695179"><img src=https://img.shields.io/badge/-failure-red></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/6005627299/job/16288695179"><img src=https://img.shields.io/badge/-failure-red></a>
|
non_process
|
fix linalg test tensorflow inv numpy a href src jax a href src tensorflow a href src torch a href src paddle a href src
| 0
|
229,201
| 18,286,657,022
|
IssuesEvent
|
2021-10-05 11:04:03
|
DILCISBoard/eark-ip-test-corpus
|
https://api.github.com/repos/DILCISBoard/eark-ip-test-corpus
|
closed
|
CSIP42 Test Case Description
|
test case corpus package
|
**Specification:**
- **Name:** E-ARK CSIP
- **Version:** 2.0-DRAFT
- **URL:** http://earkcsip.dilcis.eu/
**Requirement:**
- **Id:** CSIP42
- **Link:** http://earkcsip.dilcis.eu/#CSIP42
**Error Level:** ERROR
**Description:**
CSIP42 | File creation date amdSec/digiprovMD/mdRef/@CREATED | Date the linked file was created. | 1..1 MUST
-- | -- | -- | --
|
1.0
|
CSIP42 Test Case Description - **Specification:**
- **Name:** E-ARK CSIP
- **Version:** 2.0-DRAFT
- **URL:** http://earkcsip.dilcis.eu/
**Requirement:**
- **Id:** CSIP42
- **Link:** http://earkcsip.dilcis.eu/#CSIP42
**Error Level:** ERROR
**Description:**
CSIP42 | File creation date amdSec/digiprovMD/mdRef/@CREATED | Date the linked file was created. | 1..1 MUST
-- | -- | -- | --
|
non_process
|
test case description specification name e ark csip version draft url requirement id link error level error description file creation date amdsec digiprovmd mdref created date the linked file was created must
| 0
|
260,276
| 27,781,305,315
|
IssuesEvent
|
2023-03-16 21:19:39
|
snowdensb/caseflow
|
https://api.github.com/repos/snowdensb/caseflow
|
opened
|
CVE-2023-28155 (Medium) detected in request-2.88.2.tgz
|
Mend: dependency security vulnerability
|
## CVE-2023-28155 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>request-2.88.2.tgz</b></p></summary>
<p>Simplified HTTP request client.</p>
<p>Library home page: <a href="https://registry.npmjs.org/request/-/request-2.88.2.tgz">https://registry.npmjs.org/request/-/request-2.88.2.tgz</a></p>
<p>Path to dependency file: /client/package.json</p>
<p>Path to vulnerable library: /client/node_modules/request/package.json</p>
<p>
Dependency Hierarchy:
- jsdom-9.8.3.tgz (Root Library)
- :x: **request-2.88.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/snowdensb/caseflow/commit/81f8b3f5658022f994993a18a7653667705b7f6e">81f8b3f5658022f994993a18a7653667705b7f6e</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
** UNSUPPORTED WHEN ASSIGNED ** The Request package through 2.88.1 for Node.js allows a bypass of SSRF mitigations via an attacker-controller server that does a cross-protocol redirect (HTTP to HTTPS, or HTTPS to HTTP). NOTE: This vulnerability only affects products that are no longer supported by the maintainer.
<p>Publish Date: 2023-03-16
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-28155>CVE-2023-28155</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
|
True
|
CVE-2023-28155 (Medium) detected in request-2.88.2.tgz - ## CVE-2023-28155 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>request-2.88.2.tgz</b></p></summary>
<p>Simplified HTTP request client.</p>
<p>Library home page: <a href="https://registry.npmjs.org/request/-/request-2.88.2.tgz">https://registry.npmjs.org/request/-/request-2.88.2.tgz</a></p>
<p>Path to dependency file: /client/package.json</p>
<p>Path to vulnerable library: /client/node_modules/request/package.json</p>
<p>
Dependency Hierarchy:
- jsdom-9.8.3.tgz (Root Library)
- :x: **request-2.88.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/snowdensb/caseflow/commit/81f8b3f5658022f994993a18a7653667705b7f6e">81f8b3f5658022f994993a18a7653667705b7f6e</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
** UNSUPPORTED WHEN ASSIGNED ** The Request package through 2.88.1 for Node.js allows a bypass of SSRF mitigations via an attacker-controller server that does a cross-protocol redirect (HTTP to HTTPS, or HTTPS to HTTP). NOTE: This vulnerability only affects products that are no longer supported by the maintainer.
<p>Publish Date: 2023-03-16
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-28155>CVE-2023-28155</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
|
non_process
|
cve medium detected in request tgz cve medium severity vulnerability vulnerable library request tgz simplified http request client library home page a href path to dependency file client package json path to vulnerable library client node modules request package json dependency hierarchy jsdom tgz root library x request tgz vulnerable library found in head commit a href found in base branch master vulnerability details unsupported when assigned the request package through for node js allows a bypass of ssrf mitigations via an attacker controller server that does a cross protocol redirect http to https or https to http note this vulnerability only affects products that are no longer supported by the maintainer publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href
| 0
|
1,775
| 2,666,940,981
|
IssuesEvent
|
2015-03-22 02:35:30
|
benquarmby/jslintnet-test
|
https://api.github.com/repos/benquarmby/jslintnet-test
|
opened
|
Ability to upgrade JSLint version without recompiling
|
CodePlex
|
<b>ChrisNielsen[CodePlex]</b> <br />I am using the MSBuild task. I would like to be able to upgrade the JSLint version without needing to recompile JSLintNet. If it could read jslint.js from the file system instead of as an embedded resource, that would be great.
|
1.0
|
Ability to upgrade JSLint version without recompiling - <b>ChrisNielsen[CodePlex]</b> <br />I am using the MSBuild task. I would like to be able to upgrade the JSLint version without needing to recompile JSLintNet. If it could read jslint.js from the file system instead of as an embedded resource, that would be great.
|
non_process
|
ability to upgrade jslint version without recompiling chrisnielsen i am using the msbuild task i would like to be able to upgrade the jslint version without needing to recompile jslintnet if it could read jslint js from the file system instead of as an embedded resource that would be great
| 0
|
10,646
| 13,446,217,972
|
IssuesEvent
|
2020-09-08 12:38:58
|
MHRA/products
|
https://api.github.com/repos/MHRA/products
|
closed
|
DOCUMENTATION - PARs Process Map
|
EPIC - PARs process
|
### User want
As a internal user I would like to have access to a PARs process map, so that I can understand / amend the process in the future.
### Acceptance Criteria
**Customer acceptance criteria**
The process map follows the BPMN methodology
Agency “House” Best Practice / Style guidelines have been used
The process map plots the end to end PARs upload / amend / delete process
The process map can be imported into Visual Paradigm (XMI or XML formats - XMI standards preferred)
**Technical acceptance criteria**
**Data acceptance criteria**
**Testing acceptance criteria**
**Data - Potential impact**
**Size**
**Value**
**Effort**
### Exit Criteria met
- [ ] Backlog
- [ ] Discovery
- [ ] DUXD
- [ ] Development
- [ ] Quality Assurance
- [ ] Release and Validate
|
1.0
|
DOCUMENTATION - PARs Process Map - ### User want
As a internal user I would like to have access to a PARs process map, so that I can understand / amend the process in the future.
### Acceptance Criteria
**Customer acceptance criteria**
The process map follows the BPMN methodology
Agency “House” Best Practice / Style guidelines have been used
The process map plots the end to end PARs upload / amend / delete process
The process map can be imported into Visual Paradigm (XMI or XML formats - XMI standards preferred)
**Technical acceptance criteria**
**Data acceptance criteria**
**Testing acceptance criteria**
**Data - Potential impact**
**Size**
**Value**
**Effort**
### Exit Criteria met
- [ ] Backlog
- [ ] Discovery
- [ ] DUXD
- [ ] Development
- [ ] Quality Assurance
- [ ] Release and Validate
|
process
|
documentation pars process map user want as a internal user i would like to have access to a pars process map so that i can understand amend the process in the future acceptance criteria customer acceptance criteria the process map follows the bpmn methodology agency “house” best practice style guidelines have been used the process map plots the end to end pars upload amend delete process the process map can be imported into visual paradigm xmi or xml formats xmi standards preferred technical acceptance criteria data acceptance criteria testing acceptance criteria data potential impact size value effort exit criteria met backlog discovery duxd development quality assurance release and validate
| 1
|
304,990
| 23,093,646,524
|
IssuesEvent
|
2022-07-26 17:19:38
|
ericspring08/Awesome-Programming-Languages
|
https://api.github.com/repos/ericspring08/Awesome-Programming-Languages
|
closed
|
Add New Language: V
|
documentation
|
# Description
Simple language for building maintainable programs. You can learn the entire language by going through the [documentation](https://github.com/vlang/v/blob/master/doc/docs.md) over a weekend, and in most cases there's only one way to do something. This results in simple, readable, and maintainable code. Despite being simple, V gives a lot of power to the developer and can be used in pretty much every field, including systems programming, webdev, gamedev, GUI, mobile, science, embedded, tooling, etc. V is very similar to Go. If you know Go, you already know ≈80% of V.
|
1.0
|
Add New Language: V - # Description
Simple language for building maintainable programs. You can learn the entire language by going through the [documentation](https://github.com/vlang/v/blob/master/doc/docs.md) over a weekend, and in most cases there's only one way to do something. This results in simple, readable, and maintainable code. Despite being simple, V gives a lot of power to the developer and can be used in pretty much every field, including systems programming, webdev, gamedev, GUI, mobile, science, embedded, tooling, etc. V is very similar to Go. If you know Go, you already know ≈80% of V.
|
non_process
|
add new language v description simple language for building maintainable programs you can learn the entire language by going through the over a weekend and in most cases there s only one way to do something this results in simple readable and maintainable code despite being simple v gives a lot of power to the developer and can be used in pretty much every field including systems programming webdev gamedev gui mobile science embedded tooling etc v is very similar to go if you know go you already know ≈ of v
| 0
|
10,812
| 13,609,289,435
|
IssuesEvent
|
2020-09-23 04:50:24
|
googleapis/java-dialogflow
|
https://api.github.com/repos/googleapis/java-dialogflow
|
closed
|
Dependency Dashboard
|
api: dialogflow type: process
|
This issue contains a list of Renovate updates and their statuses.
## Open
These updates have all been created already. Click a checkbox below to force a retry/rebase of any.
- [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-dialogflow-2.x -->chore(deps): update dependency com.google.cloud:google-cloud-dialogflow to v2.2.0
- [ ] <!-- rebase-branch=renovate/com.google.cloud-libraries-bom-10.x -->chore(deps): update dependency com.google.cloud:libraries-bom to v10
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
|
1.0
|
Dependency Dashboard - This issue contains a list of Renovate updates and their statuses.
## Open
These updates have all been created already. Click a checkbox below to force a retry/rebase of any.
- [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-dialogflow-2.x -->chore(deps): update dependency com.google.cloud:google-cloud-dialogflow to v2.2.0
- [ ] <!-- rebase-branch=renovate/com.google.cloud-libraries-bom-10.x -->chore(deps): update dependency com.google.cloud:libraries-bom to v10
---
- [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
|
process
|
dependency dashboard this issue contains a list of renovate updates and their statuses open these updates have all been created already click a checkbox below to force a retry rebase of any chore deps update dependency com google cloud google cloud dialogflow to chore deps update dependency com google cloud libraries bom to check this box to trigger a request for renovate to run again on this repository
| 1
|
9,279
| 12,303,759,590
|
IssuesEvent
|
2020-05-11 19:16:15
|
nextgenhealthcare/connect
|
https://api.github.com/repos/nextgenhealthcare/connect
|
closed
|
Reprocessing a single message by pressing enter causes message to reprocess twice
|
message reprocess reprocessing twice
|
In the Channel Messages window, when you highlight a single message and then reprocess it, pressing 'Enter' to select the OK button, two new messages are sent through the channel.
Imported Issue. Original Details:
Jira Issue Key: MIRTH-2038
Reporter: narupley
Created: 2011-12-21T13:01:53.000-0800
|
2.0
|
Reprocessing a single message by pressing enter causes message to reprocess twice - In the Channel Messages window, when you highlight a single message and then reprocess it, pressing 'Enter' to select the OK button, two new messages are sent through the channel.
Imported Issue. Original Details:
Jira Issue Key: MIRTH-2038
Reporter: narupley
Created: 2011-12-21T13:01:53.000-0800
|
process
|
reprocessing a single message by pressing enter causes message to reprocess twice in the channel messages window when you highlight a single message and then reprocess it pressing enter to select the ok button two new messages are sent through the channel imported issue original details jira issue key mirth reporter narupley created
| 1
|
392,142
| 11,584,044,637
|
IssuesEvent
|
2020-02-22 15:03:03
|
IntegratedTransportPlanning/gcvt
|
https://api.github.com/repos/IntegratedTransportPlanning/gcvt
|
opened
|
Charts are broken on the web server
|
bug high priority
|
E.g. http://gcvt.cmcaine.co.uk/api/charts?scenarios=TentRail,DoNothing&variable=Ton&rows=all&width=800&height=500
```
######################### ERROR STACKTRACE #########################
IOError: could not spawn setenv(`/root/.julia/artifacts/a42f5c72a500a8683c08a5e818d27a104fc9cd5a/bin/node /root/.julia/artifacts/8e150e7c219ecfd8b18fb27ff830dc6b3628bf86/vl2vg.js`; dir="/root/.julia/artifacts/8e150e7c219ecfd8b18fb27ff830dc6b3628bf86"): no such file or directory (ENOENT)
```
Node does genuinely not exist at that location (or anywhere else) so presumably a package somewhere hasn't installed it properly.
Handy debugging step on the web server: `docker exec -it $(docker ps | grep src_julia | cut -d' ' -f1) "/bin/bash"` dumps you in a terminal in the Julia container.
|
1.0
|
Charts are broken on the web server - E.g. http://gcvt.cmcaine.co.uk/api/charts?scenarios=TentRail,DoNothing&variable=Ton&rows=all&width=800&height=500
```
######################### ERROR STACKTRACE #########################
IOError: could not spawn setenv(`/root/.julia/artifacts/a42f5c72a500a8683c08a5e818d27a104fc9cd5a/bin/node /root/.julia/artifacts/8e150e7c219ecfd8b18fb27ff830dc6b3628bf86/vl2vg.js`; dir="/root/.julia/artifacts/8e150e7c219ecfd8b18fb27ff830dc6b3628bf86"): no such file or directory (ENOENT)
```
Node does genuinely not exist at that location (or anywhere else) so presumably a package somewhere hasn't installed it properly.
Handy debugging step on the web server: `docker exec -it $(docker ps | grep src_julia | cut -d' ' -f1) "/bin/bash"` dumps you in a terminal in the Julia container.
|
non_process
|
charts are broken on the web server e g error stacktrace ioerror could not spawn setenv root julia artifacts bin node root julia artifacts js dir root julia artifacts no such file or directory enoent node does genuinely not exist at that location or anywhere else so presumably a package somewhere hasn t installed it properly handy debugging step on the web server docker exec it docker ps grep src julia cut d bin bash dumps you in a terminal in the julia container
| 0
|
95,101
| 11,954,043,811
|
IssuesEvent
|
2020-04-03 22:21:58
|
trisagion-games/Convergence-game
|
https://api.github.com/repos/trisagion-games/Convergence-game
|
closed
|
Tutorial Design (Start of Game with Console & Room 1)
|
level-design
|
Introducing idea of using a Terminal and Hacking, as well as Rooms and Keys
|
1.0
|
Tutorial Design (Start of Game with Console & Room 1) - Introducing idea of using a Terminal and Hacking, as well as Rooms and Keys
|
non_process
|
tutorial design start of game with console room introducing idea of using a terminal and hacking as well as rooms and keys
| 0
|
5,857
| 8,680,781,519
|
IssuesEvent
|
2018-12-01 14:07:08
|
bitshares/bitshares-community-ui
|
https://api.github.com/repos/bitshares/bitshares-community-ui
|
closed
|
Improvement on Login via secure key screen
|
Login P1 process
|
on the Login screen for the 'SECURE KEY' tab after user browses and selects the key file, and 'password' field appears, **hide** 'backup phrase' field.
|
1.0
|
Improvement on Login via secure key screen - on the Login screen for the 'SECURE KEY' tab after user browses and selects the key file, and 'password' field appears, **hide** 'backup phrase' field.
|
process
|
improvement on login via secure key screen on the login screen for the secure key tab after user browses and selects the key file and password field appears hide backup phrase field
| 1
|
37,538
| 8,308,927,564
|
IssuesEvent
|
2018-09-24 02:04:17
|
ESAPI/esapi-java-legacy
|
https://api.github.com/repos/ESAPI/esapi-java-legacy
|
closed
|
Remove deprecated fields in Encoder interface
|
Component-Encoder General Code Cleanup Maintainability
|
Remove all the deprecated fields in the `Encoder` interface which have equivalents in the `EncoderConstants` class. Related to closed GitHub issue #264.
Specifically, remove the following deprecated fields:
```
public final static char[] CHAR_LOWERS = EncoderConstants.CHAR_LOWERS;
public final static char[] CHAR_UPPERS = EncoderConstants.CHAR_UPPERS;
public final static char[] CHAR_DIGITS = EncoderConstants.CHAR_DIGITS;
public final static char[] CHAR_SPECIALS = EncoderConstants.CHAR_SPECIALS;
public final static char[] CHAR_LETTERS = EncoderConstants.CHAR_LETTERS;
public final static char[] CHAR_ALPHANUMERICS = EncoderConstants.CHAR_ALPHANUMERICS;
public final static char[] CHAR_PASSWORD_LOWERS = EncoderConstants.CHAR_PASSWORD_LOWERS;
public final static char[] CHAR_PASSWORD_UPPERS = EncoderConstants.CHAR_PASSWORD_UPPERS;
public final static char[] CHAR_PASSWORD_DIGITS = EncoderConstants.CHAR_PASSWORD_DIGITS;
public final static char[] CHAR_PASSWORD_SPECIALS = EncoderConstants.CHAR_PASSWORD_SPECIALS;
public final static char[] CHAR_PASSWORD_LETTERS = EncoderConstants.CHAR_PASSWORD_LETTERS;
```
|
2.0
|
Remove deprecated fields in Encoder interface - Remove all the deprecated fields in the `Encoder` interface which have equivalents in the `EncoderConstants` class. Related to closed GitHub issue #264.
Specifically, remove the following deprecated fields:
```
public final static char[] CHAR_LOWERS = EncoderConstants.CHAR_LOWERS;
public final static char[] CHAR_UPPERS = EncoderConstants.CHAR_UPPERS;
public final static char[] CHAR_DIGITS = EncoderConstants.CHAR_DIGITS;
public final static char[] CHAR_SPECIALS = EncoderConstants.CHAR_SPECIALS;
public final static char[] CHAR_LETTERS = EncoderConstants.CHAR_LETTERS;
public final static char[] CHAR_ALPHANUMERICS = EncoderConstants.CHAR_ALPHANUMERICS;
public final static char[] CHAR_PASSWORD_LOWERS = EncoderConstants.CHAR_PASSWORD_LOWERS;
public final static char[] CHAR_PASSWORD_UPPERS = EncoderConstants.CHAR_PASSWORD_UPPERS;
public final static char[] CHAR_PASSWORD_DIGITS = EncoderConstants.CHAR_PASSWORD_DIGITS;
public final static char[] CHAR_PASSWORD_SPECIALS = EncoderConstants.CHAR_PASSWORD_SPECIALS;
public final static char[] CHAR_PASSWORD_LETTERS = EncoderConstants.CHAR_PASSWORD_LETTERS;
```
|
non_process
|
remove deprecated fields in encoder interface remove all the deprecated fields in the encoder interface which have equivalents in the encoderconstants class related to closed github issue specifically remove the following deprecated fields public final static char char lowers encoderconstants char lowers public final static char char uppers encoderconstants char uppers public final static char char digits encoderconstants char digits public final static char char specials encoderconstants char specials public final static char char letters encoderconstants char letters public final static char char alphanumerics encoderconstants char alphanumerics public final static char char password lowers encoderconstants char password lowers public final static char char password uppers encoderconstants char password uppers public final static char char password digits encoderconstants char password digits public final static char char password specials encoderconstants char password specials public final static char char password letters encoderconstants char password letters
| 0
|
819,036
| 30,717,501,421
|
IssuesEvent
|
2023-07-27 13:55:20
|
IRPTeam/IRP
|
https://api.github.com/repos/IRPTeam/IRP
|
closed
|
Negative stock control
|
Priority
|
Логика проверки остатков ActualStock по каждому документу меняется с цикла с запросом остатков по каждому документу на один запрос остатков и оборотов в разрезе регистратора
При проверке остатков по FreeStock проверяем только конечный остаток (без ограничения по дате). Проверку по каждому документу убираем.
|
1.0
|
Negative stock control - Логика проверки остатков ActualStock по каждому документу меняется с цикла с запросом остатков по каждому документу на один запрос остатков и оборотов в разрезе регистратора
При проверке остатков по FreeStock проверяем только конечный остаток (без ограничения по дате). Проверку по каждому документу убираем.
|
non_process
|
negative stock control логика проверки остатков actualstock по каждому документу меняется с цикла с запросом остатков по каждому документу на один запрос остатков и оборотов в разрезе регистратора при проверке остатков по freestock проверяем только конечный остаток без ограничения по дате проверку по каждому документу убираем
| 0
|
813,469
| 30,459,073,164
|
IssuesEvent
|
2023-07-17 04:38:39
|
Haidoe/arc
|
https://api.github.com/repos/Haidoe/arc
|
opened
|
Produciton Report Forms missing in Mobile Screen
|
bug priority-medium style
|
## Bug Report
**Reporter: ❗️**
@ksdhir
**Describe the bug: ❗️**
When production report page is accessed on mobile the sidebar takes the whole page and the forms are missing.
**Steps to reproduce: ❗️**
1. Go to '[Home Page'](https://arc-app.ca/home)
2. Click on 'Any existing production'
3. Change the browser window to 'mobile view'
4. See the forms are no longer visible.
**Screenshots or Video**
***With Sidebar opened***

***With Sidebar closed***

**Expected behavior: ❗️**
The layout should look like this:

**Actual behavior: ❗️**
The sidebar takes the whole window and somehow the forms are still missing even if the sidebar is closed.
**Possible Solution:**
If you have any ideas or suggestions on how to fix the bug, please mention them here.
**Environment:**
- Device & Operating System:
- Browser and Version(if applicable):
- Jira Ticket(if applicable):
- Any other relevant information about your environment.
**Additional context:**
Add any other context about the problem here.
**Follow up checklist: ❗️**
- [ ] Add Assignee
- [ ] Label priority ( priority-low, priority-medium, priority-high )
- [ ] Label Milestone ( Alpha, Beta )
- [ ] Label Issue Type ( Style, Code, API )
|
1.0
|
Produciton Report Forms missing in Mobile Screen - ## Bug Report
**Reporter: ❗️**
@ksdhir
**Describe the bug: ❗️**
When production report page is accessed on mobile the sidebar takes the whole page and the forms are missing.
**Steps to reproduce: ❗️**
1. Go to '[Home Page'](https://arc-app.ca/home)
2. Click on 'Any existing production'
3. Change the browser window to 'mobile view'
4. See the forms are no longer visible.
**Screenshots or Video**
***With Sidebar opened***

***With Sidebar closed***

**Expected behavior: ❗️**
The layout should look like this:

**Actual behavior: ❗️**
The sidebar takes the whole window and somehow the forms are still missing even if the sidebar is closed.
**Possible Solution:**
If you have any ideas or suggestions on how to fix the bug, please mention them here.
**Environment:**
- Device & Operating System:
- Browser and Version(if applicable):
- Jira Ticket(if applicable):
- Any other relevant information about your environment.
**Additional context:**
Add any other context about the problem here.
**Follow up checklist: ❗️**
- [ ] Add Assignee
- [ ] Label priority ( priority-low, priority-medium, priority-high )
- [ ] Label Milestone ( Alpha, Beta )
- [ ] Label Issue Type ( Style, Code, API )
|
non_process
|
produciton report forms missing in mobile screen bug report reporter ❗️ ksdhir describe the bug ❗️ when production report page is accessed on mobile the sidebar takes the whole page and the forms are missing steps to reproduce ❗️ go to click on any existing production change the browser window to mobile view see the forms are no longer visible screenshots or video with sidebar opened with sidebar closed expected behavior ❗️ the layout should look like this actual behavior ❗️ the sidebar takes the whole window and somehow the forms are still missing even if the sidebar is closed possible solution if you have any ideas or suggestions on how to fix the bug please mention them here environment device operating system browser and version if applicable jira ticket if applicable any other relevant information about your environment additional context add any other context about the problem here follow up checklist ❗️ add assignee label priority priority low priority medium priority high label milestone alpha beta label issue type style code api
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.