Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
19,392
25,534,378,949
IssuesEvent
2022-11-29 10:49:35
inmanta/web-console
https://api.github.com/repos/inmanta/web-console
opened
Improved cleanup on Jenkins pipline
process
The current Jenkinsfile executes the following commands in the `Testing with cypress` stage of the pipline: ``` yarn run setup-server:lsm; yarn run cypress-test; yarn run kill-server:lsm ``` The script will stop executing as soon as one command fails. This means that when a test case fails, `yarn run kill-server:lsm` will not get executed because `yarn run cypress-test` has exited with a non-zero exit code. It would be good to move the `yarn run kill-server:lsm` command to the `cleanup` post build step of the pipeline to ensure that a cleanup happens in all circumstances.
1.0
Improved cleanup on Jenkins pipline - The current Jenkinsfile executes the following commands in the `Testing with cypress` stage of the pipline: ``` yarn run setup-server:lsm; yarn run cypress-test; yarn run kill-server:lsm ``` The script will stop executing as soon as one command fails. This means that when a test case fails, `yarn run kill-server:lsm` will not get executed because `yarn run cypress-test` has exited with a non-zero exit code. It would be good to move the `yarn run kill-server:lsm` command to the `cleanup` post build step of the pipeline to ensure that a cleanup happens in all circumstances.
process
improved cleanup on jenkins pipline the current jenkinsfile executes the following commands in the testing with cypress stage of the pipline yarn run setup server lsm yarn run cypress test yarn run kill server lsm the script will stop executing as soon as one command fails this means that when a test case fails yarn run kill server lsm will not get executed because yarn run cypress test has exited with a non zero exit code it would be good to move the yarn run kill server lsm command to the cleanup post build step of the pipeline to ensure that a cleanup happens in all circumstances
1
12,014
14,738,399,412
IssuesEvent
2021-01-07 04:39:07
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
Structured Releases
anc-process anp-important
In GitLab by @kdjstudios on May 18, 2018, 09:03 @smasih @pchaudhary @tim.traylor @thomas.gentry Hello Team, I am needing to know which server we are using for the primary releases to Production? I will be making all Help Tool Tips, News Feed, Wiki Help, and other updates to this one server then that will get released along with the other updates on our bi weekly schedule.
1.0
Structured Releases - In GitLab by @kdjstudios on May 18, 2018, 09:03 @smasih @pchaudhary @tim.traylor @thomas.gentry Hello Team, I am needing to know which server we are using for the primary releases to Production? I will be making all Help Tool Tips, News Feed, Wiki Help, and other updates to this one server then that will get released along with the other updates on our bi weekly schedule.
process
structured releases in gitlab by kdjstudios on may smasih pchaudhary tim traylor thomas gentry hello team i am needing to know which server we are using for the primary releases to production i will be making all help tool tips news feed wiki help and other updates to this one server then that will get released along with the other updates on our bi weekly schedule
1
18,783
24,690,287,548
IssuesEvent
2022-10-19 08:02:42
geneontology/go-ontology
https://api.github.com/repos/geneontology/go-ontology
closed
homeostatic process (GO:0042592) parentage query
cellular processes
I have been wondering why homeostatic process (GO:0042592) Any biological process involved in the maintenance of an internal steady state. is a descendant of regulation of biological quality (GO:0065008) Any process that modulates a qualitative or quantitative trait of a biological quality. A biological quality is a measurable attribute of an organism or part of an organism, such as size, mass, shape, color, etc. homeostasis terms are problematic anyway, because a lot of regulation *is*. homeostasis, but also not all of homeostasis is regulation. So we probably need to really think about what we are trying to capture with homeostasis. For example it might have more value for metabolic processes, but it would be very difficult to be precise about when a pathway is part of homeostasis. Int the example of. "zinc transport" some transporters transport since whenever they detect it, but some pathways are activated to maintain the correct concentrations in organelles (as far as I understand it). In the short term, so that we can properly assess which genes are annotated to "regulation of process" terms, it might be useful to remove the "regulation" parent from cellular homeostasis?
1.0
homeostatic process (GO:0042592) parentage query - I have been wondering why homeostatic process (GO:0042592) Any biological process involved in the maintenance of an internal steady state. is a descendant of regulation of biological quality (GO:0065008) Any process that modulates a qualitative or quantitative trait of a biological quality. A biological quality is a measurable attribute of an organism or part of an organism, such as size, mass, shape, color, etc. homeostasis terms are problematic anyway, because a lot of regulation *is*. homeostasis, but also not all of homeostasis is regulation. So we probably need to really think about what we are trying to capture with homeostasis. For example it might have more value for metabolic processes, but it would be very difficult to be precise about when a pathway is part of homeostasis. Int the example of. "zinc transport" some transporters transport since whenever they detect it, but some pathways are activated to maintain the correct concentrations in organelles (as far as I understand it). In the short term, so that we can properly assess which genes are annotated to "regulation of process" terms, it might be useful to remove the "regulation" parent from cellular homeostasis?
process
homeostatic process go parentage query i have been wondering why homeostatic process go any biological process involved in the maintenance of an internal steady state is a descendant of regulation of biological quality go any process that modulates a qualitative or quantitative trait of a biological quality a biological quality is a measurable attribute of an organism or part of an organism such as size mass shape color etc homeostasis terms are problematic anyway because a lot of regulation is homeostasis but also not all of homeostasis is regulation so we probably need to really think about what we are trying to capture with homeostasis for example it might have more value for metabolic processes but it would be very difficult to be precise about when a pathway is part of homeostasis int the example of zinc transport some transporters transport since whenever they detect it but some pathways are activated to maintain the correct concentrations in organelles as far as i understand it in the short term so that we can properly assess which genes are annotated to regulation of process terms it might be useful to remove the regulation parent from cellular homeostasis
1
5,648
8,513,547,837
IssuesEvent
2018-10-31 16:18:07
FaruqRahmadani/Presensi-Guru-2
https://api.github.com/repos/FaruqRahmadani/Presensi-Guru-2
closed
Variabel sass useless
PROCESS
### File : resources/assets/sass/_mentah/_variable.scss > ![image](https://user-images.githubusercontent.com/11961119/47668324-1c150980-dbe3-11e8-99fe-4c64f975a68e.png) ### File :resources/assets/sass/_mentah/_login.scss > ![image](https://user-images.githubusercontent.com/11961119/47668390-41a21300-dbe3-11e8-98a5-897c6c0034b3.png) ### Permasalahan asaku sayang mun kita sdh declare variabel tp kd dipakai.
1.0
Variabel sass useless - ### File : resources/assets/sass/_mentah/_variable.scss > ![image](https://user-images.githubusercontent.com/11961119/47668324-1c150980-dbe3-11e8-99fe-4c64f975a68e.png) ### File :resources/assets/sass/_mentah/_login.scss > ![image](https://user-images.githubusercontent.com/11961119/47668390-41a21300-dbe3-11e8-98a5-897c6c0034b3.png) ### Permasalahan asaku sayang mun kita sdh declare variabel tp kd dipakai.
process
variabel sass useless file resources assets sass mentah variable scss file resources assets sass mentah login scss permasalahan asaku sayang mun kita sdh declare variabel tp kd dipakai
1
6,099
8,958,406,426
IssuesEvent
2019-01-27 14:05:01
lxqt/lxqt
https://api.github.com/repos/lxqt/lxqt
closed
Notification icon doesn't update
wont-process-this wontfix
When an application shows notifications the first image is used for all said app's notifications. This is a problem when something like spotify is showing what song is being played, as it will only show the album art of the first song played for all future songs - presumably it's cached on first display. Version: 0.13
1.0
Notification icon doesn't update - When an application shows notifications the first image is used for all said app's notifications. This is a problem when something like spotify is showing what song is being played, as it will only show the album art of the first song played for all future songs - presumably it's cached on first display. Version: 0.13
process
notification icon doesn t update when an application shows notifications the first image is used for all said app s notifications this is a problem when something like spotify is showing what song is being played as it will only show the album art of the first song played for all future songs presumably it s cached on first display version
1
17,961
23,970,061,130
IssuesEvent
2022-09-13 06:55:56
cypress-io/cypress
https://api.github.com/repos/cypress-io/cypress
opened
[ACI] Investigate better testing strategy
process: tests stage: internal
### What would you like? Testing ACI PRs is [hard and time consuming](https://github.com/cypress-io/cypress/pull/23735#issuecomment-1243296033), since a lot of the logic is "after X days, Y should happen" which is based on data in the global cache (read from user's machine). See this PR https://github.com/cypress-io/cypress/pull/23735 for an example of the complexity to test manually. I think we can automate most of this. The gaps in our current infra (within the monorepo) are: 1. no examples on how to update the cache (the file on disc persisted between app sessions) 2. (probably) no endpoint to refetch this (I think we just fetch it once, on startup) If we could implement these things, we could probably automate about 90% of the test plan in the PR linked above. We should do this. ### Why is this needed? _No response_ ### Other _No response_
1.0
[ACI] Investigate better testing strategy - ### What would you like? Testing ACI PRs is [hard and time consuming](https://github.com/cypress-io/cypress/pull/23735#issuecomment-1243296033), since a lot of the logic is "after X days, Y should happen" which is based on data in the global cache (read from user's machine). See this PR https://github.com/cypress-io/cypress/pull/23735 for an example of the complexity to test manually. I think we can automate most of this. The gaps in our current infra (within the monorepo) are: 1. no examples on how to update the cache (the file on disc persisted between app sessions) 2. (probably) no endpoint to refetch this (I think we just fetch it once, on startup) If we could implement these things, we could probably automate about 90% of the test plan in the PR linked above. We should do this. ### Why is this needed? _No response_ ### Other _No response_
process
investigate better testing strategy what would you like testing aci prs is since a lot of the logic is after x days y should happen which is based on data in the global cache read from user s machine see this pr for an example of the complexity to test manually i think we can automate most of this the gaps in our current infra within the monorepo are no examples on how to update the cache the file on disc persisted between app sessions probably no endpoint to refetch this i think we just fetch it once on startup if we could implement these things we could probably automate about of the test plan in the pr linked above we should do this why is this needed no response other no response
1
11,157
13,957,693,713
IssuesEvent
2020-10-24 08:11:03
alexanderkotsev/geoportal
https://api.github.com/repos/alexanderkotsev/geoportal
opened
BE: Missing resources in Geoportal
BE - Belgium Geoportal Harvesting process
Collected from the Geoportal Workshop online survey answers: At Walloon Region level: - Geoportal detects the link to VS and DS correclty. - Regarding the unique resource identifier, Geoportal doesn’t detect the code and Namespace declared in the metadata cause to the fact we declare two Unique ID: i) the global ID using gmd:citation/gmd:CI_Citation/gmd:identifier/*/gmd:code element and its sub-type “RS_Identifier” which is linked to the dataset and ii) the local ID using gmd:citation/gmd:CI_Citation/gmd:identifier/*/gmd:code element and its sub-type “MD_Identifier” which is a characterString referring the short name of the dataset in our internal database. http://inspire-geoportal.ec.europa.eu/download_details.html?view=downloadDetails&resourceId=%2FINSPIRE-f5201eb0-2404-11e5-8130-52540004b857_20181206-123520%2Fservices%2F1%2FPullResults%2F141-160%2Fseries%2F7&expandedSection=metadata
1.0
BE: Missing resources in Geoportal - Collected from the Geoportal Workshop online survey answers: At Walloon Region level: - Geoportal detects the link to VS and DS correclty. - Regarding the unique resource identifier, Geoportal doesn’t detect the code and Namespace declared in the metadata cause to the fact we declare two Unique ID: i) the global ID using gmd:citation/gmd:CI_Citation/gmd:identifier/*/gmd:code element and its sub-type “RS_Identifier” which is linked to the dataset and ii) the local ID using gmd:citation/gmd:CI_Citation/gmd:identifier/*/gmd:code element and its sub-type “MD_Identifier” which is a characterString referring the short name of the dataset in our internal database. http://inspire-geoportal.ec.europa.eu/download_details.html?view=downloadDetails&resourceId=%2FINSPIRE-f5201eb0-2404-11e5-8130-52540004b857_20181206-123520%2Fservices%2F1%2FPullResults%2F141-160%2Fseries%2F7&expandedSection=metadata
process
be missing resources in geoportal collected from the geoportal workshop online survey answers at walloon region level geoportal detects the link to vs and ds correclty regarding the unique resource identifier geoportal doesn rsquo t detect the code and namespace declared in the metadata cause to the fact we declare two unique id i the global id using gmd citation gmd ci citation gmd identifier gmd code element and its sub type ldquo rs identifier rdquo which is linked to the dataset and ii the local id using gmd citation gmd ci citation gmd identifier gmd code element and its sub type ldquo md identifier rdquo which is a characterstring referring the short name of the dataset in our internal database
1
12,360
14,888,288,181
IssuesEvent
2021-01-20 19:37:09
encode/uvicorn
https://api.github.com/repos/encode/uvicorn
closed
Huge number of workers fired up when I tried to run the app using PyInstaller generated byte code
multiprocessing need confirmation
### Checklist <!-- Please make sure you check all these items before submitting your bug report. --> - [Yes ] The bug is reproducible against the latest release and/or `master`. [Tested using uvicorn version 0.12.1] - [Yes] There are no similar issues or pull requests to fix it yet. ### Describe the bug <!-- A clear and concise description of what the bug is. --> We bundled our app using pyinstaller byte code. When we try to use 4 workers to run the application, a huge number of workers are fired up resulting in our server going down. When we remove the "workers = 4" parameter in the Uvicorn.run(), everything is working fine. ### To reproduce <!-- Provide a *minimal* example with steps to reproduce the bug locally. NOTE: try to keep any external dependencies *at an absolute minimum* . In other words, remove anything that doesn't make the bug go away. --> Server: RHEL Server on AWS EC2 instance When I generated the pyinstaller byte code using the below files, everything works fine. "test.py" and "test.spec" files used to generate the byte code executable. Steps: 1. Install PyInstaller and run "pyinstaller --onefile -w test.spec". A byte code executable 'test' is generated in dist/ directory. 2. Run the executable as './test'. [testspec.txt](https://github.com/encode/uvicorn/files/5839125/testspec.txt) [testpy.txt](https://github.com/encode/uvicorn/files/5839126/testpy.txt) <img width="560" alt="Screen Shot 2021-01-19 at 5 58 39 PM" src="https://user-images.githubusercontent.com/32053777/105105227-a9253100-5a81-11eb-884c-fa6fb1f187de.png"> But when we set the number of workers to 4, huge number of workers are fired up resulting in server going down. [test1spec.txt](https://github.com/encode/uvicorn/files/5839127/test1spec.txt) [test1py.txt](https://github.com/encode/uvicorn/files/5839128/test1py.txt) Steps: Repeat the same procedure as above with test1.spec. <img width="537" alt="Screen Shot 2021-01-19 at 5 45 52 PM" src="https://user-images.githubusercontent.com/32053777/105105312-deca1a00-5a81-11eb-91db-e9b344902393.png"> ### Expected behavior <!-- A clear and concise description of what you expected to happen. --> Expected 4 workers to start. ### Actual behavior <!-- A clear and concise description of what actually happens. --> Huge number of workers are fired up resulting in server going down. ### Debugging material <!-- Any tracebacks, screenshots, etc. that can help understanding the problem. NOTE: - Please list tracebacks in full (don't truncate them). - If relevant, consider turning on DEBUG or TRACE logs for additional details (see the Logging section on https://www.uvicorn.org/settings/ specifically the `log-level` flag). - Consider using `<details>` to make tracebacks/logs collapsible if they're very large (see https://gist.github.com/ericclemmons/b146fe5da72ca1f706b2ef72a20ac39d). --> ### Environment OS: RHEL Server on AWS EC2 instance Python: Python 3.6.8 Uvicorn version: uvicorn 0.12.1 with CPython 3.8.3 on Linux - uvicorn.run('test:app', host='0.0.0.0', workers=4, port=7500) ### Additional context <!-- Any additional information that can help understanding the problem. Eg. linked issues, or a description of what you were trying to achieve. -->
1.0
Huge number of workers fired up when I tried to run the app using PyInstaller generated byte code - ### Checklist <!-- Please make sure you check all these items before submitting your bug report. --> - [Yes ] The bug is reproducible against the latest release and/or `master`. [Tested using uvicorn version 0.12.1] - [Yes] There are no similar issues or pull requests to fix it yet. ### Describe the bug <!-- A clear and concise description of what the bug is. --> We bundled our app using pyinstaller byte code. When we try to use 4 workers to run the application, a huge number of workers are fired up resulting in our server going down. When we remove the "workers = 4" parameter in the Uvicorn.run(), everything is working fine. ### To reproduce <!-- Provide a *minimal* example with steps to reproduce the bug locally. NOTE: try to keep any external dependencies *at an absolute minimum* . In other words, remove anything that doesn't make the bug go away. --> Server: RHEL Server on AWS EC2 instance When I generated the pyinstaller byte code using the below files, everything works fine. "test.py" and "test.spec" files used to generate the byte code executable. Steps: 1. Install PyInstaller and run "pyinstaller --onefile -w test.spec". A byte code executable 'test' is generated in dist/ directory. 2. Run the executable as './test'. [testspec.txt](https://github.com/encode/uvicorn/files/5839125/testspec.txt) [testpy.txt](https://github.com/encode/uvicorn/files/5839126/testpy.txt) <img width="560" alt="Screen Shot 2021-01-19 at 5 58 39 PM" src="https://user-images.githubusercontent.com/32053777/105105227-a9253100-5a81-11eb-884c-fa6fb1f187de.png"> But when we set the number of workers to 4, huge number of workers are fired up resulting in server going down. [test1spec.txt](https://github.com/encode/uvicorn/files/5839127/test1spec.txt) [test1py.txt](https://github.com/encode/uvicorn/files/5839128/test1py.txt) Steps: Repeat the same procedure as above with test1.spec. <img width="537" alt="Screen Shot 2021-01-19 at 5 45 52 PM" src="https://user-images.githubusercontent.com/32053777/105105312-deca1a00-5a81-11eb-91db-e9b344902393.png"> ### Expected behavior <!-- A clear and concise description of what you expected to happen. --> Expected 4 workers to start. ### Actual behavior <!-- A clear and concise description of what actually happens. --> Huge number of workers are fired up resulting in server going down. ### Debugging material <!-- Any tracebacks, screenshots, etc. that can help understanding the problem. NOTE: - Please list tracebacks in full (don't truncate them). - If relevant, consider turning on DEBUG or TRACE logs for additional details (see the Logging section on https://www.uvicorn.org/settings/ specifically the `log-level` flag). - Consider using `<details>` to make tracebacks/logs collapsible if they're very large (see https://gist.github.com/ericclemmons/b146fe5da72ca1f706b2ef72a20ac39d). --> ### Environment OS: RHEL Server on AWS EC2 instance Python: Python 3.6.8 Uvicorn version: uvicorn 0.12.1 with CPython 3.8.3 on Linux - uvicorn.run('test:app', host='0.0.0.0', workers=4, port=7500) ### Additional context <!-- Any additional information that can help understanding the problem. Eg. linked issues, or a description of what you were trying to achieve. -->
process
huge number of workers fired up when i tried to run the app using pyinstaller generated byte code checklist the bug is reproducible against the latest release and or master there are no similar issues or pull requests to fix it yet describe the bug we bundled our app using pyinstaller byte code when we try to use workers to run the application a huge number of workers are fired up resulting in our server going down when we remove the workers parameter in the uvicorn run everything is working fine to reproduce provide a minimal example with steps to reproduce the bug locally note try to keep any external dependencies at an absolute minimum in other words remove anything that doesn t make the bug go away server rhel server on aws instance when i generated the pyinstaller byte code using the below files everything works fine test py and test spec files used to generate the byte code executable steps install pyinstaller and run pyinstaller onefile w test spec a byte code executable test is generated in dist directory run the executable as test img width alt screen shot at pm src but when we set the number of workers to huge number of workers are fired up resulting in server going down steps repeat the same procedure as above with spec img width alt screen shot at pm src expected behavior expected workers to start actual behavior huge number of workers are fired up resulting in server going down debugging material any tracebacks screenshots etc that can help understanding the problem note please list tracebacks in full don t truncate them if relevant consider turning on debug or trace logs for additional details see the logging section on specifically the log level flag consider using to make tracebacks logs collapsible if they re very large see environment os rhel server on aws instance python python uvicorn version uvicorn with cpython on linux uvicorn run test app host workers port additional context any additional information that can help understanding the problem eg linked issues or a description of what you were trying to achieve
1
86,953
17,108,289,047
IssuesEvent
2021-07-09 22:04:39
phetsims/circuit-construction-kit-common
https://api.github.com/repos/phetsims/circuit-construction-kit-common
closed
What is the convention for naming imported sounds?
dev:code-review
In code review https://github.com/phetsims/circuit-construction-kit-common/issues/672 @jonathanolson asked: ```js //REVIEW: What is the convention for naming of these? import dogBarkSound from '../../sounds/dog-bark_mp3.js'; ```
1.0
What is the convention for naming imported sounds? - In code review https://github.com/phetsims/circuit-construction-kit-common/issues/672 @jonathanolson asked: ```js //REVIEW: What is the convention for naming of these? import dogBarkSound from '../../sounds/dog-bark_mp3.js'; ```
non_process
what is the convention for naming imported sounds in code review jonathanolson asked js review what is the convention for naming of these import dogbarksound from sounds dog bark js
0
20,413
27,072,268,080
IssuesEvent
2023-02-14 08:01:47
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
Open source internal version of BlazeJavacMainTest
P4 type: process team-Rules-Java stale
After https://bazel-review.git.corp.google.com/c/10751/ is submitted there will be 2 versions of this test (internal + external), which is not desirable. Open sourcing the internal version also improves contributions test coverage.
1.0
Open source internal version of BlazeJavacMainTest - After https://bazel-review.git.corp.google.com/c/10751/ is submitted there will be 2 versions of this test (internal + external), which is not desirable. Open sourcing the internal version also improves contributions test coverage.
process
open source internal version of blazejavacmaintest after is submitted there will be versions of this test internal external which is not desirable open sourcing the internal version also improves contributions test coverage
1
176,476
14,583,616,295
IssuesEvent
2020-12-18 14:16:05
ChainYo/StableMatching
https://api.github.com/repos/ChainYo/StableMatching
closed
Update Readme
documentation good first issue
We need to update [Readme.md] to keep informations up to date. - [x] Quick presentation for the project - [x] Explanation of the goal of the project - [x] For who it's designed for - [x] Explanations or a Link to know how it's work (probably a link to docs) - [x] Link for contact - [x] How to contribute
1.0
Update Readme - We need to update [Readme.md] to keep informations up to date. - [x] Quick presentation for the project - [x] Explanation of the goal of the project - [x] For who it's designed for - [x] Explanations or a Link to know how it's work (probably a link to docs) - [x] Link for contact - [x] How to contribute
non_process
update readme we need to update to keep informations up to date quick presentation for the project explanation of the goal of the project for who it s designed for explanations or a link to know how it s work probably a link to docs link for contact how to contribute
0
287,577
21,660,642,673
IssuesEvent
2022-05-06 18:40:02
dpaceoffice/ECommerce
https://api.github.com/repos/dpaceoffice/ECommerce
opened
Server Design
documentation
#7 Create a design document that fully defines the app's backend services. This document should include: complete REST API, identify all server-side logic & data processing responsibilities, define all backend tooling and libraries.
1.0
Server Design - #7 Create a design document that fully defines the app's backend services. This document should include: complete REST API, identify all server-side logic & data processing responsibilities, define all backend tooling and libraries.
non_process
server design create a design document that fully defines the app s backend services this document should include complete rest api identify all server side logic data processing responsibilities define all backend tooling and libraries
0
13,882
16,654,734,355
IssuesEvent
2021-06-05 10:06:11
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[PM] Responsive issue in Sites tab > UI issues
Bug P2 Participant manager Process: Fixed Process: Tested dev
Responsive issue in Sites tab > UI issues 1. Add site button is wrapping up with the above text ![mbs4](https://user-images.githubusercontent.com/71445210/115548120-b7732b00-a2c4-11eb-80a0-2838c83c2ddc.png) 2. There should be a space between images and text and other UI issues ![mb5](https://user-images.githubusercontent.com/71445210/115548256-e25d7f00-a2c4-11eb-8d64-29300f451f21.png)
2.0
[PM] Responsive issue in Sites tab > UI issues - Responsive issue in Sites tab > UI issues 1. Add site button is wrapping up with the above text ![mbs4](https://user-images.githubusercontent.com/71445210/115548120-b7732b00-a2c4-11eb-80a0-2838c83c2ddc.png) 2. There should be a space between images and text and other UI issues ![mb5](https://user-images.githubusercontent.com/71445210/115548256-e25d7f00-a2c4-11eb-8d64-29300f451f21.png)
process
responsive issue in sites tab ui issues responsive issue in sites tab ui issues add site button is wrapping up with the above text there should be a space between images and text and other ui issues
1
153,864
13,529,051,284
IssuesEvent
2020-09-15 17:40:16
aws/chalice
https://api.github.com/repos/aws/chalice
closed
chalice deploy import error, with layers
documentation
Hi I got module not found error when deploying, I specify the function to use layers `{ "version": "2.0", "app_name": "myapp", "stages": { "dev": { "api_gateway_stage": "api", "manage_iam_role": false, "iam_role_arn": "arn", "lambda_functions": { "myfunc": { "layers": ["layer1", "layer2"] } } } }, }` I did not specify the module in requirements.txt since I have the layers specified. I probably missing something, could not find much docs on layers for chalice. Thanks Dong
1.0
chalice deploy import error, with layers - Hi I got module not found error when deploying, I specify the function to use layers `{ "version": "2.0", "app_name": "myapp", "stages": { "dev": { "api_gateway_stage": "api", "manage_iam_role": false, "iam_role_arn": "arn", "lambda_functions": { "myfunc": { "layers": ["layer1", "layer2"] } } } }, }` I did not specify the module in requirements.txt since I have the layers specified. I probably missing something, could not find much docs on layers for chalice. Thanks Dong
non_process
chalice deploy import error with layers hi i got module not found error when deploying i specify the function to use layers version app name myapp stages dev api gateway stage api manage iam role false iam role arn arn lambda functions myfunc layers i did not specify the module in requirements txt since i have the layers specified i probably missing something could not find much docs on layers for chalice thanks dong
0
97,350
28,209,870,970
IssuesEvent
2023-04-05 02:36:13
opensearch-project/performance-analyzer
https://api.github.com/repos/opensearch-project/performance-analyzer
closed
performance-analyzer.log is not rotated
bug Build Libraries & Interfaces
**Describe the bug** We are running a three node cluster on Docker and our monitoring system reported that the node disks were running out of space. After digging out we have noticed that `/usr/share/opensearch/logs/performance-analyzer.log` was not rotated (the file was huge, several GiB). I went through the documentation and found nothing on the subject. I have to run this command once in a while on each Opensearch node to workaround the issue: ``` docker exec <opensearch_container> truncate -s0 /usr/share/opensearch/logs/performance-analyzer.log ``` **To Reproduce** Steps to reproduce the behavior: 1. Deploy an Opensearch on Docker 2. Let it run for a while **Expected behavior** `/usr/share/opensearch/logs/performance-analyzer.log` should be rotated **Plugins** Security plugin **Host/Environment:** - OS: Debian Bullseye - Opensearch Docker image opensearchproject/opensearch:1.2.4 - Docker CE 20.10.12
1.0
performance-analyzer.log is not rotated - **Describe the bug** We are running a three node cluster on Docker and our monitoring system reported that the node disks were running out of space. After digging out we have noticed that `/usr/share/opensearch/logs/performance-analyzer.log` was not rotated (the file was huge, several GiB). I went through the documentation and found nothing on the subject. I have to run this command once in a while on each Opensearch node to workaround the issue: ``` docker exec <opensearch_container> truncate -s0 /usr/share/opensearch/logs/performance-analyzer.log ``` **To Reproduce** Steps to reproduce the behavior: 1. Deploy an Opensearch on Docker 2. Let it run for a while **Expected behavior** `/usr/share/opensearch/logs/performance-analyzer.log` should be rotated **Plugins** Security plugin **Host/Environment:** - OS: Debian Bullseye - Opensearch Docker image opensearchproject/opensearch:1.2.4 - Docker CE 20.10.12
non_process
performance analyzer log is not rotated describe the bug we are running a three node cluster on docker and our monitoring system reported that the node disks were running out of space after digging out we have noticed that usr share opensearch logs performance analyzer log was not rotated the file was huge several gib i went through the documentation and found nothing on the subject i have to run this command once in a while on each opensearch node to workaround the issue docker exec truncate usr share opensearch logs performance analyzer log to reproduce steps to reproduce the behavior deploy an opensearch on docker let it run for a while expected behavior usr share opensearch logs performance analyzer log should be rotated plugins security plugin host environment os debian bullseye opensearch docker image opensearchproject opensearch docker ce
0
6,216
9,126,231,862
IssuesEvent
2019-02-24 19:59:41
rtcharity/eahub.org
https://api.github.com/repos/rtcharity/eahub.org
opened
Periodic check-ins?
Process
I'm of two minds as to whether we should restart doing periodic check-ins of what people are working on. On the one hand, it's useful for me to know what everyone else is doing so that I can factor that into what I'm doing, e.g., avoid duplicating work. And I'm sure the product and project managers would like to know this too. On the other hand they mean committing to deadlines for short-term tasks much more frequently (since people are counting on you doing what you said you'd do), which I'm a bit reluctant to do since I'm doing all of this in what spare time I can manage. And I don't want to wind up flaking again. What do other people think?
1.0
Periodic check-ins? - I'm of two minds as to whether we should restart doing periodic check-ins of what people are working on. On the one hand, it's useful for me to know what everyone else is doing so that I can factor that into what I'm doing, e.g., avoid duplicating work. And I'm sure the product and project managers would like to know this too. On the other hand they mean committing to deadlines for short-term tasks much more frequently (since people are counting on you doing what you said you'd do), which I'm a bit reluctant to do since I'm doing all of this in what spare time I can manage. And I don't want to wind up flaking again. What do other people think?
process
periodic check ins i m of two minds as to whether we should restart doing periodic check ins of what people are working on on the one hand it s useful for me to know what everyone else is doing so that i can factor that into what i m doing e g avoid duplicating work and i m sure the product and project managers would like to know this too on the other hand they mean committing to deadlines for short term tasks much more frequently since people are counting on you doing what you said you d do which i m a bit reluctant to do since i m doing all of this in what spare time i can manage and i don t want to wind up flaking again what do other people think
1
20,594
27,263,788,878
IssuesEvent
2023-02-22 16:33:17
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
Why can't we reference template files from an artifact?
devops/prod doc-bug Pri1 devops-cicd-process/tech
This section of the document: https://docs.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops#insert-a-template States that "Template files need to exist on your filesystem at the start of a pipeline run. You can't reference templates in an artifact." If we want to reuse a common repo for release template steps, it is important that we use a snapshot version of the templates steps at the time of build. That way the release steps are known and predictable throughout all stages of a release. Referencing release steps based on a repo filesystem path is dangerous because changes to that template for an in-flight release could break or introduce unwanted changes. Can you please clarify if there is any workaround to reference release templates from the current build artifact? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 6724abea-bbdc-bf66-ed5e-3214fa6c3e66 * Version Independent ID: 4f8dab21-3f0e-da32-cc0e-1d85c13c0065 * Content: [Templates - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops#step-re-use) * Content Source: [docs/pipelines/process/templates.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/templates.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
Why can't we reference template files from an artifact? - This section of the document: https://docs.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops#insert-a-template States that "Template files need to exist on your filesystem at the start of a pipeline run. You can't reference templates in an artifact." If we want to reuse a common repo for release template steps, it is important that we use a snapshot version of the templates steps at the time of build. That way the release steps are known and predictable throughout all stages of a release. Referencing release steps based on a repo filesystem path is dangerous because changes to that template for an in-flight release could break or introduce unwanted changes. Can you please clarify if there is any workaround to reference release templates from the current build artifact? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 6724abea-bbdc-bf66-ed5e-3214fa6c3e66 * Version Independent ID: 4f8dab21-3f0e-da32-cc0e-1d85c13c0065 * Content: [Templates - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/templates?view=azure-devops#step-re-use) * Content Source: [docs/pipelines/process/templates.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/templates.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
why can t we reference template files from an artifact this section of the document states that template files need to exist on your filesystem at the start of a pipeline run you can t reference templates in an artifact if we want to reuse a common repo for release template steps it is important that we use a snapshot version of the templates steps at the time of build that way the release steps are known and predictable throughout all stages of a release referencing release steps based on a repo filesystem path is dangerous because changes to that template for an in flight release could break or introduce unwanted changes can you please clarify if there is any workaround to reference release templates from the current build artifact document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id bbdc version independent id content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
79,286
28,086,167,588
IssuesEvent
2023-03-30 09:54:57
snowplow/snowplow-javascript-tracker
https://api.github.com/repos/snowplow/snowplow-javascript-tracker
closed
Fix top level domain grouping initial cookie set attempt
type:defect
**Describe the bug** The algorithm behind the `discoverRootDomain` functionality checks if the domain is valid to set a cookie at the highest level domain. This algorithm currently starts from a eTLD which will not be valid in any case, e.g. `.com`, `.io` etc. This behaviour will create a console warning every time in some browsers e.g. Firefox ![image](https://user-images.githubusercontent.com/15251081/222681262-45046b16-b482-4dcc-9c97-b126ad8c63f7.png) The fix would just be to start trying from the TLD and not the eTLD. **Note** Even with the fix, if we have an eTLD with more than one part e.g. co.uk, then the warning will still be there. This happens with other trackers as well e.g. analytics.js
1.0
Fix top level domain grouping initial cookie set attempt - **Describe the bug** The algorithm behind the `discoverRootDomain` functionality checks if the domain is valid to set a cookie at the highest level domain. This algorithm currently starts from a eTLD which will not be valid in any case, e.g. `.com`, `.io` etc. This behaviour will create a console warning every time in some browsers e.g. Firefox ![image](https://user-images.githubusercontent.com/15251081/222681262-45046b16-b482-4dcc-9c97-b126ad8c63f7.png) The fix would just be to start trying from the TLD and not the eTLD. **Note** Even with the fix, if we have an eTLD with more than one part e.g. co.uk, then the warning will still be there. This happens with other trackers as well e.g. analytics.js
non_process
fix top level domain grouping initial cookie set attempt describe the bug the algorithm behind the discoverrootdomain functionality checks if the domain is valid to set a cookie at the highest level domain this algorithm currently starts from a etld which will not be valid in any case e g com io etc this behaviour will create a console warning every time in some browsers e g firefox the fix would just be to start trying from the tld and not the etld note even with the fix if we have an etld with more than one part e g co uk then the warning will still be there this happens with other trackers as well e g analytics js
0
14,329
17,362,631,044
IssuesEvent
2021-07-29 23:43:48
googleapis/python-spanner
https://api.github.com/repos/googleapis/python-spanner
closed
tests.system.test_system.TestSessionAPI: many tests failed
api: spanner flakybot: issue type: process
Many tests failed at the same time in this package. * I will close this issue when there are no more failures in this package _and_ there is at least one pass. * No new issues will be filed for this package until this issue is closed. * If there are already issues for individual test cases, I will close them when the corresponding test passes. You can close them earlier, if you prefer, and I won't reopen them while this issue is still open. Here are the tests that failed: * test_batch_insert_or_update_then_query * test_batch_insert_then_read * test_batch_insert_then_read_all_datatypes * test_batch_insert_then_read_string_array_of_string * test_batch_insert_w_commit_timestamp * test_empty_read * test_empty_read_with_single_keys_index * test_execute_partitioned_dml * test_execute_sql_returning_array_of_struct * test_execute_sql_returning_empty_array_of_struct * test_execute_sql_returning_transfinite_floats * test_execute_sql_select_1 * test_execute_sql_w_bool_bindings * test_execute_sql_w_bytes_bindings * test_execute_sql_w_date_bindings * test_execute_sql_w_float64_bindings * test_execute_sql_w_float_bindings_transfinite * test_execute_sql_w_int64_bindings * test_execute_sql_w_manual_consume * test_execute_sql_w_numeric_bindings * test_execute_sql_w_query_param_struct * test_execute_sql_w_string_bindings * test_execute_sql_w_timestamp_bindings * test_invalid_type * test_multiuse_snapshot_execute_sql_isolation_strong * test_multiuse_snapshot_read_isolation_exact_staleness * test_multiuse_snapshot_read_isolation_read_timestamp * test_multiuse_snapshot_read_isolation_strong * test_partition_query * test_partition_read_w_index * test_read_partial_range_from_beginning * test_read_partial_range_until_end * test_read_w_index * test_read_w_limit * test_read_w_multiple_keys * test_read_w_ranges * test_read_w_single_key * test_read_with_multiple_keys_index * test_read_with_range_keys_and_index_closed_closed * test_read_with_range_keys_and_index_closed_open * test_read_with_range_keys_and_index_open_closed * test_read_with_range_keys_and_index_open_open * test_read_with_range_keys_index_closed_closed * test_read_with_range_keys_index_closed_open * test_read_with_range_keys_index_limit_closed_closed * test_read_with_range_keys_index_limit_closed_open * test_read_with_range_keys_index_limit_open_closed * test_read_with_range_keys_index_limit_open_open * test_read_with_range_keys_index_open_closed * test_read_with_range_keys_index_open_open * test_read_with_range_keys_index_single_key * test_read_with_single_keys_index * test_session_crud * test_snapshot_read_w_various_staleness * test_transaction_batch_update_and_execute_dml * test_transaction_batch_update_success * test_transaction_batch_update_w_parent_span * test_transaction_batch_update_w_syntax_error * test_transaction_batch_update_wo_statements * test_transaction_execute_sql_w_dml_read_rollback * test_transaction_execute_update_read_commit * test_transaction_execute_update_then_insert_commit * test_transaction_query_w_concurrent_updates * test_transaction_read_and_insert_or_update_then_commit * test_transaction_read_and_insert_then_exception * test_transaction_read_and_insert_then_rollback * test_transaction_read_w_abort * test_transaction_read_w_concurrent_updates ----- commit: 2487800e31842a44dcc37937c325e130c8c926b0 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/306a2e02-87cb-4be9-be31-37456ec7a8a2), [Sponge](http://sponge2/306a2e02-87cb-4be9-be31-37456ec7a8a2) status: failed
1.0
tests.system.test_system.TestSessionAPI: many tests failed - Many tests failed at the same time in this package. * I will close this issue when there are no more failures in this package _and_ there is at least one pass. * No new issues will be filed for this package until this issue is closed. * If there are already issues for individual test cases, I will close them when the corresponding test passes. You can close them earlier, if you prefer, and I won't reopen them while this issue is still open. Here are the tests that failed: * test_batch_insert_or_update_then_query * test_batch_insert_then_read * test_batch_insert_then_read_all_datatypes * test_batch_insert_then_read_string_array_of_string * test_batch_insert_w_commit_timestamp * test_empty_read * test_empty_read_with_single_keys_index * test_execute_partitioned_dml * test_execute_sql_returning_array_of_struct * test_execute_sql_returning_empty_array_of_struct * test_execute_sql_returning_transfinite_floats * test_execute_sql_select_1 * test_execute_sql_w_bool_bindings * test_execute_sql_w_bytes_bindings * test_execute_sql_w_date_bindings * test_execute_sql_w_float64_bindings * test_execute_sql_w_float_bindings_transfinite * test_execute_sql_w_int64_bindings * test_execute_sql_w_manual_consume * test_execute_sql_w_numeric_bindings * test_execute_sql_w_query_param_struct * test_execute_sql_w_string_bindings * test_execute_sql_w_timestamp_bindings * test_invalid_type * test_multiuse_snapshot_execute_sql_isolation_strong * test_multiuse_snapshot_read_isolation_exact_staleness * test_multiuse_snapshot_read_isolation_read_timestamp * test_multiuse_snapshot_read_isolation_strong * test_partition_query * test_partition_read_w_index * test_read_partial_range_from_beginning * test_read_partial_range_until_end * test_read_w_index * test_read_w_limit * test_read_w_multiple_keys * test_read_w_ranges * test_read_w_single_key * test_read_with_multiple_keys_index * test_read_with_range_keys_and_index_closed_closed * test_read_with_range_keys_and_index_closed_open * test_read_with_range_keys_and_index_open_closed * test_read_with_range_keys_and_index_open_open * test_read_with_range_keys_index_closed_closed * test_read_with_range_keys_index_closed_open * test_read_with_range_keys_index_limit_closed_closed * test_read_with_range_keys_index_limit_closed_open * test_read_with_range_keys_index_limit_open_closed * test_read_with_range_keys_index_limit_open_open * test_read_with_range_keys_index_open_closed * test_read_with_range_keys_index_open_open * test_read_with_range_keys_index_single_key * test_read_with_single_keys_index * test_session_crud * test_snapshot_read_w_various_staleness * test_transaction_batch_update_and_execute_dml * test_transaction_batch_update_success * test_transaction_batch_update_w_parent_span * test_transaction_batch_update_w_syntax_error * test_transaction_batch_update_wo_statements * test_transaction_execute_sql_w_dml_read_rollback * test_transaction_execute_update_read_commit * test_transaction_execute_update_then_insert_commit * test_transaction_query_w_concurrent_updates * test_transaction_read_and_insert_or_update_then_commit * test_transaction_read_and_insert_then_exception * test_transaction_read_and_insert_then_rollback * test_transaction_read_w_abort * test_transaction_read_w_concurrent_updates ----- commit: 2487800e31842a44dcc37937c325e130c8c926b0 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/306a2e02-87cb-4be9-be31-37456ec7a8a2), [Sponge](http://sponge2/306a2e02-87cb-4be9-be31-37456ec7a8a2) status: failed
process
tests system test system testsessionapi many tests failed many tests failed at the same time in this package i will close this issue when there are no more failures in this package and there is at least one pass no new issues will be filed for this package until this issue is closed if there are already issues for individual test cases i will close them when the corresponding test passes you can close them earlier if you prefer and i won t reopen them while this issue is still open here are the tests that failed test batch insert or update then query test batch insert then read test batch insert then read all datatypes test batch insert then read string array of string test batch insert w commit timestamp test empty read test empty read with single keys index test execute partitioned dml test execute sql returning array of struct test execute sql returning empty array of struct test execute sql returning transfinite floats test execute sql select test execute sql w bool bindings test execute sql w bytes bindings test execute sql w date bindings test execute sql w bindings test execute sql w float bindings transfinite test execute sql w bindings test execute sql w manual consume test execute sql w numeric bindings test execute sql w query param struct test execute sql w string bindings test execute sql w timestamp bindings test invalid type test multiuse snapshot execute sql isolation strong test multiuse snapshot read isolation exact staleness test multiuse snapshot read isolation read timestamp test multiuse snapshot read isolation strong test partition query test partition read w index test read partial range from beginning test read partial range until end test read w index test read w limit test read w multiple keys test read w ranges test read w single key test read with multiple keys index test read with range keys and index closed closed test read with range keys and index closed open test read with range keys and index open closed test read with range keys and index open open test read with range keys index closed closed test read with range keys index closed open test read with range keys index limit closed closed test read with range keys index limit closed open test read with range keys index limit open closed test read with range keys index limit open open test read with range keys index open closed test read with range keys index open open test read with range keys index single key test read with single keys index test session crud test snapshot read w various staleness test transaction batch update and execute dml test transaction batch update success test transaction batch update w parent span test transaction batch update w syntax error test transaction batch update wo statements test transaction execute sql w dml read rollback test transaction execute update read commit test transaction execute update then insert commit test transaction query w concurrent updates test transaction read and insert or update then commit test transaction read and insert then exception test transaction read and insert then rollback test transaction read w abort test transaction read w concurrent updates commit buildurl status failed
1
79,015
7,692,912,230
IssuesEvent
2018-05-18 00:13:55
tnibert/project-camelot
https://api.github.com/repos/tnibert/project-camelot
closed
Make username login case insensitive
expandtests
- Change username in user login and registration to convert to lower case - Go through existing db and verify case, change if necessary
1.0
Make username login case insensitive - - Change username in user login and registration to convert to lower case - Go through existing db and verify case, change if necessary
non_process
make username login case insensitive change username in user login and registration to convert to lower case go through existing db and verify case change if necessary
0
586,047
17,568,669,458
IssuesEvent
2021-08-14 07:56:44
xnd-project/rbc
https://api.github.com/repos/xnd-project/rbc
opened
rbc 0.6.0 release plan
high-priority
For rbc 0.6.0, the following issues/PRs must be resolved/landed: - [ ] https://github.com/omnisci/omniscidb-internal/pull/5765 - required to support `set_output_row_size` within rbc - [ ] https://github.com/xnd-project/rbc/pull/341 - implements `set_output_row_size` support in rbc - [ ] Optionally https://github.com/xnd-project/rbc/pull/353 - implements numba 0.54 support
1.0
rbc 0.6.0 release plan - For rbc 0.6.0, the following issues/PRs must be resolved/landed: - [ ] https://github.com/omnisci/omniscidb-internal/pull/5765 - required to support `set_output_row_size` within rbc - [ ] https://github.com/xnd-project/rbc/pull/341 - implements `set_output_row_size` support in rbc - [ ] Optionally https://github.com/xnd-project/rbc/pull/353 - implements numba 0.54 support
non_process
rbc release plan for rbc the following issues prs must be resolved landed required to support set output row size within rbc implements set output row size support in rbc optionally implements numba support
0
13,696
16,454,792,657
IssuesEvent
2021-05-21 10:59:28
ORNL-AMO/AMO-Tools-Desktop
https://api.github.com/repos/ORNL-AMO/AMO-Tools-Desktop
opened
Adding PHAST material updates options in modification
Process Heating
In some process heating calculators adding a new material/gas/solid should also update the options for the modification side. check PHAST modify-conditions also.
1.0
Adding PHAST material updates options in modification - In some process heating calculators adding a new material/gas/solid should also update the options for the modification side. check PHAST modify-conditions also.
process
adding phast material updates options in modification in some process heating calculators adding a new material gas solid should also update the options for the modification side check phast modify conditions also
1
308,552
26,611,490,856
IssuesEvent
2023-01-24 00:49:06
OvercastCommunity/CommunityMaps
https://api.github.com/repos/OvercastCommunity/CommunityMaps
opened
[Payload Race/10v10] TF2: Pipeline
map submission contest
### Checklist - [X] I have [pruned](https://pgm.dev/docs/guides/packaging/pruning-chunks) the map. - [X] I have agreed with assigning the CC BY-SA license to this map, as mentioned in the [README](https://github.com/OvercastCommunity/CommunityMaps#readme). - [X] I have created an XML file. - [X] I have created a map image. - [X] I have uploaded the map zip file to a file sharing service. - [X] The map has been tested locally to make sure it works. ### Map Name TF2: Pipeline ### Gamemode & Map Description _TF2: Pipeline_ is a highly experimental Conquest map, using an alternate form of the new Payload gamemode. Instead of using Attack/Defense, this map is a Payload Race wherein each team has their own payload, and much attempt to attack and defend at the same time. As a proof of concept, this is an adaptation of the first stage of the map Pipeline from Team Fortress 2, with some creative liberties with gameplay and aesthetics in mind. The map has some cover for bow fire and also multiple paths to encourage melee flanking as well. With the current xml setup, it's possible to stop the enemy team's cart movement by entering its range, carts take 2 minutes of nonstop movement to hit the enemy end zone, and carts will slowly lose progress with no teammates within its range. There are alternates to all of these, so they can be changed in the future beyond the contest if necessary. ### Team Sizes 10v10 ### Screenshots https://imgur.com/a/SNoaSrq ### XML https://gist.github.com/Antnecb/2d45436b14d67f531ab6406a0c20e217 ### Map Image ![map](https://user-images.githubusercontent.com/26776253/214188724-5e5384b7-df27-42b4-a008-285d29061f50.png) ### Map Download https://drive.google.com/file/d/1UOyvwSJI7MoCbwp8OmmLjApInBq-V_BX/view?usp=sharing
1.0
[Payload Race/10v10] TF2: Pipeline - ### Checklist - [X] I have [pruned](https://pgm.dev/docs/guides/packaging/pruning-chunks) the map. - [X] I have agreed with assigning the CC BY-SA license to this map, as mentioned in the [README](https://github.com/OvercastCommunity/CommunityMaps#readme). - [X] I have created an XML file. - [X] I have created a map image. - [X] I have uploaded the map zip file to a file sharing service. - [X] The map has been tested locally to make sure it works. ### Map Name TF2: Pipeline ### Gamemode & Map Description _TF2: Pipeline_ is a highly experimental Conquest map, using an alternate form of the new Payload gamemode. Instead of using Attack/Defense, this map is a Payload Race wherein each team has their own payload, and much attempt to attack and defend at the same time. As a proof of concept, this is an adaptation of the first stage of the map Pipeline from Team Fortress 2, with some creative liberties with gameplay and aesthetics in mind. The map has some cover for bow fire and also multiple paths to encourage melee flanking as well. With the current xml setup, it's possible to stop the enemy team's cart movement by entering its range, carts take 2 minutes of nonstop movement to hit the enemy end zone, and carts will slowly lose progress with no teammates within its range. There are alternates to all of these, so they can be changed in the future beyond the contest if necessary. ### Team Sizes 10v10 ### Screenshots https://imgur.com/a/SNoaSrq ### XML https://gist.github.com/Antnecb/2d45436b14d67f531ab6406a0c20e217 ### Map Image ![map](https://user-images.githubusercontent.com/26776253/214188724-5e5384b7-df27-42b4-a008-285d29061f50.png) ### Map Download https://drive.google.com/file/d/1UOyvwSJI7MoCbwp8OmmLjApInBq-V_BX/view?usp=sharing
non_process
pipeline checklist i have the map i have agreed with assigning the cc by sa license to this map as mentioned in the i have created an xml file i have created a map image i have uploaded the map zip file to a file sharing service the map has been tested locally to make sure it works map name pipeline gamemode map description pipeline is a highly experimental conquest map using an alternate form of the new payload gamemode instead of using attack defense this map is a payload race wherein each team has their own payload and much attempt to attack and defend at the same time as a proof of concept this is an adaptation of the first stage of the map pipeline from team fortress with some creative liberties with gameplay and aesthetics in mind the map has some cover for bow fire and also multiple paths to encourage melee flanking as well with the current xml setup it s possible to stop the enemy team s cart movement by entering its range carts take minutes of nonstop movement to hit the enemy end zone and carts will slowly lose progress with no teammates within its range there are alternates to all of these so they can be changed in the future beyond the contest if necessary team sizes screenshots xml map image map download
0
1,058
3,526,135,905
IssuesEvent
2016-01-14 02:22:20
osresearch/vst
https://api.github.com/repos/osresearch/vst
opened
VstBuffer should track segments
processing
The `VstBuffer` class should track segments, rather than individual points. This might make the sort algorithm simpler and maps well to the `line()` override.
1.0
VstBuffer should track segments - The `VstBuffer` class should track segments, rather than individual points. This might make the sort algorithm simpler and maps well to the `line()` override.
process
vstbuffer should track segments the vstbuffer class should track segments rather than individual points this might make the sort algorithm simpler and maps well to the line override
1
22,367
31,116,577,417
IssuesEvent
2023-08-15 00:35:32
hsmusic/hsmusic-wiki
https://api.github.com/repos/hsmusic/hsmusic-wiki
opened
Remove "Has Track Art" field from albums & ensure track covers are always associated with credits
scope: data processing scope: data format - YAML scope: data format - thing objects
Direct quote from Discord #code-quarantine as it covers the situation sufficiently: --- ```js // Now this is a doozy! static hasUniqueCoverArt( track, albumData, coverArtistContribsByRef, hasCoverArt ) { if (coverArtistContribsByRef?.length > 0) { return true; } else if (coverArtistContribsByRef) { return false; } else if (hasCoverArt === false) { return false; } else if (Track.findAlbum(track, albumData)?.hasTrackArt) { return true; } else { return false; } } ``` understanding: 1. if `coverArtistContribsByRef` is non-null (`?.`) and contains at least one item (`.length > 0`), return true 2. if `coverArtistContribsByRef` is non-null (truthy) and contains zero items (by implication), return false 3. if `hasCoverArt` is explicitly set to false for this track, return false 4. if hasTrackArt is true for the album - which it is, unless explicitly set to false - return true 5. return false do not understand why pt 2 exists at all. data files never set a field to an empty array. this is possible (they are normal YAML files so `Cover Artists: []` is valid), but never happens in practice pt 1 returns true in the correct case, but it is only relevant in the case where the album is `Has Track Art: false` else, listing no cover artists is tantamount to the wiki giving up and saying "OK, well, here is this track's art, but who knows who made it" ¯\_(ツ)_/¯ proposed simplified logic: 1. if `coverArtistContribsByRef` is non-null and contains at least one item, return true (use `empty()`) 2. if `hasCoverArt` is explicitly set to false for this track, return false 3. if the album's `trackCoverArtistContribsByRef` is non-null and contains at least one item, return true 4. return false pt 1 is the end-all case. if cover artist credits are present, then the track has cover art. period. pt 2 exists solely to override pt 3. if pt 3 is irrelevant then `hasCoverArt` should not be set to false, because pt 3 would fall through to pt 4, return false. pt 3 exists to ensure tracks automatically have unique artwork if a `Default Track Cover Artists` field has been set. (again, this can be overridden by pt 2 if select tracks do not have cover artworks while the rest share the same cover artists.) this altogether nullifies a need for the `Has Track Art` field (on albums). also, it guarantees that all track artworks are associated with cover artist credits. both "return true" cases are directly the result of cover artist contributions being present. (either directly on the track, or inherited from the album)
1.0
Remove "Has Track Art" field from albums & ensure track covers are always associated with credits - Direct quote from Discord #code-quarantine as it covers the situation sufficiently: --- ```js // Now this is a doozy! static hasUniqueCoverArt( track, albumData, coverArtistContribsByRef, hasCoverArt ) { if (coverArtistContribsByRef?.length > 0) { return true; } else if (coverArtistContribsByRef) { return false; } else if (hasCoverArt === false) { return false; } else if (Track.findAlbum(track, albumData)?.hasTrackArt) { return true; } else { return false; } } ``` understanding: 1. if `coverArtistContribsByRef` is non-null (`?.`) and contains at least one item (`.length > 0`), return true 2. if `coverArtistContribsByRef` is non-null (truthy) and contains zero items (by implication), return false 3. if `hasCoverArt` is explicitly set to false for this track, return false 4. if hasTrackArt is true for the album - which it is, unless explicitly set to false - return true 5. return false do not understand why pt 2 exists at all. data files never set a field to an empty array. this is possible (they are normal YAML files so `Cover Artists: []` is valid), but never happens in practice pt 1 returns true in the correct case, but it is only relevant in the case where the album is `Has Track Art: false` else, listing no cover artists is tantamount to the wiki giving up and saying "OK, well, here is this track's art, but who knows who made it" ¯\_(ツ)_/¯ proposed simplified logic: 1. if `coverArtistContribsByRef` is non-null and contains at least one item, return true (use `empty()`) 2. if `hasCoverArt` is explicitly set to false for this track, return false 3. if the album's `trackCoverArtistContribsByRef` is non-null and contains at least one item, return true 4. return false pt 1 is the end-all case. if cover artist credits are present, then the track has cover art. period. pt 2 exists solely to override pt 3. if pt 3 is irrelevant then `hasCoverArt` should not be set to false, because pt 3 would fall through to pt 4, return false. pt 3 exists to ensure tracks automatically have unique artwork if a `Default Track Cover Artists` field has been set. (again, this can be overridden by pt 2 if select tracks do not have cover artworks while the rest share the same cover artists.) this altogether nullifies a need for the `Has Track Art` field (on albums). also, it guarantees that all track artworks are associated with cover artist credits. both "return true" cases are directly the result of cover artist contributions being present. (either directly on the track, or inherited from the album)
process
remove has track art field from albums ensure track covers are always associated with credits direct quote from discord code quarantine as it covers the situation sufficiently js now this is a doozy static hasuniquecoverart track albumdata coverartistcontribsbyref hascoverart if coverartistcontribsbyref length return true else if coverartistcontribsbyref return false else if hascoverart false return false else if track findalbum track albumdata hastrackart return true else return false understanding if coverartistcontribsbyref is non null and contains at least one item length return true if coverartistcontribsbyref is non null truthy and contains zero items by implication return false if hascoverart is explicitly set to false for this track return false if hastrackart is true for the album which it is unless explicitly set to false return true return false do not understand why pt exists at all data files never set a field to an empty array this is possible they are normal yaml files so cover artists is valid but never happens in practice pt returns true in the correct case but it is only relevant in the case where the album is has track art false else listing no cover artists is tantamount to the wiki giving up and saying ok well here is this track s art but who knows who made it ¯ ツ ¯ proposed simplified logic if coverartistcontribsbyref is non null and contains at least one item return true use empty if hascoverart is explicitly set to false for this track return false if the album s trackcoverartistcontribsbyref is non null and contains at least one item return true return false pt is the end all case if cover artist credits are present then the track has cover art period pt exists solely to override pt if pt is irrelevant then hascoverart should not be set to false because pt would fall through to pt return false pt exists to ensure tracks automatically have unique artwork if a default track cover artists field has been set again this can be overridden by pt if select tracks do not have cover artworks while the rest share the same cover artists this altogether nullifies a need for the has track art field on albums also it guarantees that all track artworks are associated with cover artist credits both return true cases are directly the result of cover artist contributions being present either directly on the track or inherited from the album
1
15,416
19,604,716,312
IssuesEvent
2022-01-06 07:53:55
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
GDAL Buffer Vectors; not possible to provide negative buffer distance
Processing Bug
### What is the bug or the crash? The `Buffer Distance` field in `Toolbox -> GDAL -> Buffer Vectors` cannot be set to a negative value in the UI. However the tool itself (`ogr2ogr`) IS able to handle negative values properly. ### Steps to reproduce the issue * Open `Toolbox -> GDAL -> Buffer Vectors` * Enter negative value in `Buffer Distance` field * Observe it being set to 0.000 ### Versions QGIS version 3.22.2-Białowieża QGIS code revision 1601ec46d0 Qt version 5.15.2 Python version 3.9.5 GDAL/OGR version 3.4.0 PROJ version 8.2.0 EPSG Registry database version v10.038 (2021-10-21) GEOS version 3.10.0-CAPI-1.16.0 SQLite version 3.35.2 PDAL version 2.3.0 PostgreSQL client version 13.0 SpatiaLite version 5.0.1 QWT version 6.1.3 QScintilla2 version 2.11.5 OS version Windows 10 Version 2009 Active Python plugins BGTImport 3.16 inspireNL 2.6 QuickOSM 2.0.0 quick_map_services 0.19.27 db_manager 0.1.20 grassprovider 2.12.99 MetaSearch 0.3.5 processing 2.12.99 sagaprovider 2.12.99 ### Supported QGIS version - [ ] I'm running a supported QGIS version according to the roadmap. ### New profile - [ ] I tried with a new QGIS profile ### Additional context _No response_
1.0
GDAL Buffer Vectors; not possible to provide negative buffer distance - ### What is the bug or the crash? The `Buffer Distance` field in `Toolbox -> GDAL -> Buffer Vectors` cannot be set to a negative value in the UI. However the tool itself (`ogr2ogr`) IS able to handle negative values properly. ### Steps to reproduce the issue * Open `Toolbox -> GDAL -> Buffer Vectors` * Enter negative value in `Buffer Distance` field * Observe it being set to 0.000 ### Versions QGIS version 3.22.2-Białowieża QGIS code revision 1601ec46d0 Qt version 5.15.2 Python version 3.9.5 GDAL/OGR version 3.4.0 PROJ version 8.2.0 EPSG Registry database version v10.038 (2021-10-21) GEOS version 3.10.0-CAPI-1.16.0 SQLite version 3.35.2 PDAL version 2.3.0 PostgreSQL client version 13.0 SpatiaLite version 5.0.1 QWT version 6.1.3 QScintilla2 version 2.11.5 OS version Windows 10 Version 2009 Active Python plugins BGTImport 3.16 inspireNL 2.6 QuickOSM 2.0.0 quick_map_services 0.19.27 db_manager 0.1.20 grassprovider 2.12.99 MetaSearch 0.3.5 processing 2.12.99 sagaprovider 2.12.99 ### Supported QGIS version - [ ] I'm running a supported QGIS version according to the roadmap. ### New profile - [ ] I tried with a new QGIS profile ### Additional context _No response_
process
gdal buffer vectors not possible to provide negative buffer distance what is the bug or the crash the buffer distance field in toolbox gdal buffer vectors cannot be set to a negative value in the ui however the tool itself is able to handle negative values properly steps to reproduce the issue open toolbox gdal buffer vectors enter negative value in buffer distance field observe it being set to versions qgis version białowieża qgis code revision qt version python version gdal ogr version proj version epsg registry database version geos version capi sqlite version pdal version postgresql client version spatialite version qwt version version os version windows version active python plugins bgtimport inspirenl quickosm quick map services db manager grassprovider metasearch processing sagaprovider supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context no response
1
102,289
31,885,091,692
IssuesEvent
2023-09-16 21:15:48
moby/moby
https://api.github.com/repos/moby/moby
closed
No way to use both --cache-from and a local store?
area/builder area/distribution version/17.03 area/builder/classic-builder
Using the new --cache-from is a huge improvement most of the time, but it seems to cause docker to avoid looking in its, for lack of a better term, local cache, meaning that a second build of the same image starts from scratch, or at least from where the remote cache diverges. https://github.com/docker/docker/blob/master/daemon/cache.go#L12 seems to confirm that this is very much intended. It would be great to look at the local cache first and only hit the --cache-from images if it doesn't find anything. Or possibly have a way of specifying the local cache as an option, say `docker build --cache-from . myimage:v1.0` It would also be great to document this behavior as part of the `--cache-from` docs. Right now it says "Images to consider as cache sources" but it's unclear that that causes the normal cache to be unconsidered. Thanks! **Steps to reproduce the issue:** ``` docker pull myimage:v1.0 docker build --cache-from myimage:v1.0 -t myimage:v1.1 . docker build --cache-from myimage:v1.0 -t myimage:v1.1 . ``` **Describe the results you received:** The second build starts from the latest step it shares with myimage:v1.0 **Describe the results you expected:** The second build re-uses the entire cache from the first build. **Additional information you deem important (e.g. issue happens only occasionally):** **Output of `docker version`:** ``` Client: Version: 17.03.1-ce API version: 1.27 Go version: go1.7.5 Git commit: c6d412e Built: Fri Mar 24 00:45:26 2017 OS/Arch: linux/amd64 Server: Version: 17.03.1-ce API version: 1.27 (minimum version 1.12) Go version: go1.7.5 Git commit: c6d412e Built: Fri Mar 24 00:45:26 2017 OS/Arch: linux/amd64 Experimental: false ``` **Output of `docker info`:** ``` Containers: 12 Running: 0 Paused: 0 Stopped: 12 Images: 239 Server Version: 17.03.1-ce Storage Driver: overlay2 Backing Filesystem: extfs Supports d_type: true Native Overlay Diff: true Logging Driver: json-file Cgroup Driver: cgroupfs Plugins: Volume: local Network: bridge host macvlan null overlay Swarm: inactive Runtimes: runc Default Runtime: runc Init Binary: docker-init containerd version: 4ab9917febca54791c5f071a9d1f404867857fcc runc version: 54296cf40ad8143b62dbcaa1d90e520a2136ddfe init version: 949e6fa Security Options: apparmor seccomp Profile: default Kernel Version: 4.4.0-72-generic Operating System: Ubuntu 16.04.2 LTS OSType: linux Architecture: x86_64 CPUs: 2 Total Memory: 7.795 GiB Name: ip-172-31-50-48 ID: H5BG:BAQV:VVQ2:BGUO:5MQG:OXKU:T6MH:ZDXF:ZMXT:TP7G:B6K3:JIZG Docker Root Dir: /var/lib/docker Debug Mode (client): false Debug Mode (server): false Username: <redacted> Registry: https://index.docker.io/v1/ WARNING: No swap limit support Experimental: false Insecure Registries: 127.0.0.0/8 Live Restore Enabled: false ``` **Additional environment details (AWS, VirtualBox, physical, etc.):**
3.0
No way to use both --cache-from and a local store? - Using the new --cache-from is a huge improvement most of the time, but it seems to cause docker to avoid looking in its, for lack of a better term, local cache, meaning that a second build of the same image starts from scratch, or at least from where the remote cache diverges. https://github.com/docker/docker/blob/master/daemon/cache.go#L12 seems to confirm that this is very much intended. It would be great to look at the local cache first and only hit the --cache-from images if it doesn't find anything. Or possibly have a way of specifying the local cache as an option, say `docker build --cache-from . myimage:v1.0` It would also be great to document this behavior as part of the `--cache-from` docs. Right now it says "Images to consider as cache sources" but it's unclear that that causes the normal cache to be unconsidered. Thanks! **Steps to reproduce the issue:** ``` docker pull myimage:v1.0 docker build --cache-from myimage:v1.0 -t myimage:v1.1 . docker build --cache-from myimage:v1.0 -t myimage:v1.1 . ``` **Describe the results you received:** The second build starts from the latest step it shares with myimage:v1.0 **Describe the results you expected:** The second build re-uses the entire cache from the first build. **Additional information you deem important (e.g. issue happens only occasionally):** **Output of `docker version`:** ``` Client: Version: 17.03.1-ce API version: 1.27 Go version: go1.7.5 Git commit: c6d412e Built: Fri Mar 24 00:45:26 2017 OS/Arch: linux/amd64 Server: Version: 17.03.1-ce API version: 1.27 (minimum version 1.12) Go version: go1.7.5 Git commit: c6d412e Built: Fri Mar 24 00:45:26 2017 OS/Arch: linux/amd64 Experimental: false ``` **Output of `docker info`:** ``` Containers: 12 Running: 0 Paused: 0 Stopped: 12 Images: 239 Server Version: 17.03.1-ce Storage Driver: overlay2 Backing Filesystem: extfs Supports d_type: true Native Overlay Diff: true Logging Driver: json-file Cgroup Driver: cgroupfs Plugins: Volume: local Network: bridge host macvlan null overlay Swarm: inactive Runtimes: runc Default Runtime: runc Init Binary: docker-init containerd version: 4ab9917febca54791c5f071a9d1f404867857fcc runc version: 54296cf40ad8143b62dbcaa1d90e520a2136ddfe init version: 949e6fa Security Options: apparmor seccomp Profile: default Kernel Version: 4.4.0-72-generic Operating System: Ubuntu 16.04.2 LTS OSType: linux Architecture: x86_64 CPUs: 2 Total Memory: 7.795 GiB Name: ip-172-31-50-48 ID: H5BG:BAQV:VVQ2:BGUO:5MQG:OXKU:T6MH:ZDXF:ZMXT:TP7G:B6K3:JIZG Docker Root Dir: /var/lib/docker Debug Mode (client): false Debug Mode (server): false Username: <redacted> Registry: https://index.docker.io/v1/ WARNING: No swap limit support Experimental: false Insecure Registries: 127.0.0.0/8 Live Restore Enabled: false ``` **Additional environment details (AWS, VirtualBox, physical, etc.):**
non_process
no way to use both cache from and a local store using the new cache from is a huge improvement most of the time but it seems to cause docker to avoid looking in its for lack of a better term local cache meaning that a second build of the same image starts from scratch or at least from where the remote cache diverges seems to confirm that this is very much intended it would be great to look at the local cache first and only hit the cache from images if it doesn t find anything or possibly have a way of specifying the local cache as an option say docker build cache from myimage it would also be great to document this behavior as part of the cache from docs right now it says images to consider as cache sources but it s unclear that that causes the normal cache to be unconsidered thanks steps to reproduce the issue docker pull myimage docker build cache from myimage t myimage docker build cache from myimage t myimage describe the results you received the second build starts from the latest step it shares with myimage describe the results you expected the second build re uses the entire cache from the first build additional information you deem important e g issue happens only occasionally output of docker version client version ce api version go version git commit built fri mar os arch linux server version ce api version minimum version go version git commit built fri mar os arch linux experimental false output of docker info containers running paused stopped images server version ce storage driver backing filesystem extfs supports d type true native overlay diff true logging driver json file cgroup driver cgroupfs plugins volume local network bridge host macvlan null overlay swarm inactive runtimes runc default runtime runc init binary docker init containerd version runc version init version security options apparmor seccomp profile default kernel version generic operating system ubuntu lts ostype linux architecture cpus total memory gib name ip id baqv bguo oxku zdxf zmxt jizg docker root dir var lib docker debug mode client false debug mode server false username registry warning no swap limit support experimental false insecure registries live restore enabled false additional environment details aws virtualbox physical etc
0
4,325
7,236,639,196
IssuesEvent
2018-02-13 08:04:53
muflihun/residue
https://api.github.com/repos/muflihun/residue
opened
Deadlock detected in develop
area: log-processing type: bug
``` 19:03:02,606 [LogHandler] [vTRACE] [session.cc:86] Adding bytes 19:03:02,608 [LogDispatcher] [vDEBUG] [log-request-handler.cc:88] Pausing schedule for client integrity 19:03:02,609 [LogHandler] [vTRACE] [session.cc:75] Received: 522 bytes 19:03:02,609 [LogHandler] [vTRACE] [session.cc:102] Read bytes: 0fe53a0a15ad3b19847a4e5ee30bdcda:muflihun00102030:sHbsBEN2KhHtBETNCpQDU3wHPw27YHBzztBj7yckU2sHE7lLyUw35wCpxM0qALezZcvXC51YAc75cm60NMY1+mJzr4gpHWlSduK4nj8Y87aPa3/GPoBOjpIl1bTYoUghZ6lCyNgBgNmw7I8Av6oxnaiWJ6AsZARX2Oy1K/DMv5kE33XjxEVkLwRSoCGOS71dSo6LtFzF16CzGXqr/bKHB3vBv+xb8Ap9rN2w36dXGk02ydyE5EF18Pt5pLVzEIyhROlG2ErWShBcWdigYy6pY7rl/qlY/9v1GbggQ0n4rUpPjKtS50H+Z4IQ8ZVT63FMEvb6FY6hs654hx7kdssWvA4jwa7r5x0OVr2+lme8qptSpxEJ15eesTYZvgQRXTImMrGXsqo3VXQZ/AQZ6ZSNaW4/99x17SP6SYiVxNIwQgv5WAnh7j+jAO5I7zTMY66kJQtrg7tozUAIBcM8oQed9Q== [size: 522] 19:03:02,609 [LogHandler] [vDEBUG] [session.cc:171] Sending {r:0} 19:03:02,609 [LogHandler] [vTRACE] [session.cc:86] Adding bytes 19:03:02,609 [LogDispatcher] [vCRAZY] [log-request-handler.cc:100] -----============= [ BEGIN ] =============----- 19:03:02,610 [LogDispatcher] [vDEBUG] [request-handler.h:144] Raw request: 938d714753441d7cdf1167023aa7f3f9:muflihun00102030:xLRbv36YW//gqY9gVrsCOBuwoE5NM+A8iJjoE7W1Jq1/PUXOvuCk0lVvlkmGZT2Hq0Vhj3vuzdSWxqiUH2Iu3I5ufo1EbWuIvQb04eZ0k6tZxJ1wRDjhdfIaCUab9Av0nlYohOWi56pD2suVynM4xucvXuFh6mm2T+qeTyF0i7LuZybJ5bouOHlxLzSCTiI+RQNPJqQo/TyCxim5Swa/PNZwD/S9/pPoPY42XDfZehan0r/kDP9kdQw2SPTaBxG79jyqF4g61o0fsnqtXtrVVVtxgbi5ySvSy40TpdYTbG7DME9hfRlGdxciyewftqijfaQePgJweYwTUYxVtl4uttiaPbELbe1OyC6qKvn7drfgFFtIttbUBa8lf8GV8FJ7vmpMr8CZMYk/8V4PYX9kRhiSKgCHK2vUTXlCja7CbYzoTgplL5iwrQf4LNJ/6FdckPS+ZtVHl9knQokhTyGbqw== 19:03:02,610 [LogDispatcher] [vDEBUG] [request-handler.cc:58] Client: muflihun00102030 19:03:02,610 [LogDispatcher] [vCRAZY] [request-handler.cc:59] IV: 938d714753441d7cdf1167023aa7f3f9 19:03:02,610 [LogDispatcher] [vCRAZY] [request-handler.cc:66] Data (base64): xLRbv36YW//gqY9gVrsCOBuwoE5NM+A8iJjoE7W1Jq1/PUXOvuCk0lVvlkmGZT2Hq0Vhj3vuzdSWxqiUH2Iu3I5ufo1EbWuIvQb04eZ0k6tZxJ1wRDjhdfIaCUab9Av0nlYohOWi56pD2suVynM4xucvXuFh6mm2T+qeTyF0i7LuZybJ5bouOHlxLzSCTiI+RQNPJqQo/TyCxim5Swa/PNZwD/S9/pPoPY42XDfZehan0r/kDP9kdQw2SPTaBxG79jyqF4g61o0fsnqtXtrVVVtxgbi5ySvSy40TpdYTbG7DME9hfRlGdxciyewftqijfaQePgJweYwTUYxVtl4uttiaPbELbe1OyC6qKvn7drfgFFtIttbUBa8lf8GV8FJ7vmpMr8CZMYk/8V4PYX9kRhiSKgCHK2vUTXlCja7CbYzoTgplL5iwrQf4LNJ/6FdckPS+ZtVHl9knQokhTyGbqw== 19:03:02,610 [LogDispatcher] [vCRAZY] [request-handler.cc:69] Ripe command: echo 938d714753441d7cdf1167023aa7f3f9:muflihun00102030:xLRbv36YW//gqY9gVrsCOBuwoE5NM+A8iJjoE7W1Jq1/PUXOvuCk0lVvlkmGZT2Hq0Vhj3vuzdSWxqiUH2Iu3I5ufo1EbWuIvQb04eZ0k6tZxJ1wRDjhdfIaCUab9Av0nlYohOWi56pD2suVynM4xucvXuFh6mm2T+qeTyF0i7LuZybJ5bouOHlxLzSCTiI+RQNPJqQo/TyCxim5Swa/PNZwD/S9/pPoPY42XDfZehan0r/kDP9kdQw2SPTaBxG79jyqF4g61o0fsnqtXtrVVVtxgbi5ySvSy40TpdYTbG7DME9hfRlGdxciyewftqijfaQePgJweYwTUYxVtl4uttiaPbELbe1OyC6qKvn7drfgFFtIttbUBa8lf8GV8FJ7vmpMr8CZMYk/8V4PYX9kRhiSKgCHK2vUTXlCja7CbYzoTgplL5iwrQf4LNJ/6FdckPS+ZtVHl9knQokhTyGbqw== | ripe -d --aes --key 00098C6A684633C0087D9C76B47D2A57 --base64 19:03:02,610 [LogDispatcher] [vCRAZY] [request-handler.cc:76] Plain request: eNrtz7FOwzAQxvFXiTxBleA0UdSUmZWtTIjBOEcaNU6D7XRBvDsn9SUY/tvdZ/vOv/cfM7gseQpinvfdvu/q/tg3dV2X5muaNTT2LUlM1n1uYUsufNsXudkoaRo2qfy6Wg3XWZIdJDt9MlQ+uIvYFL116/rkvdFZ2+J11rTkIrhpedCiLPzZxWK3e9TzWW4y6w+aXutp0b1Ne9DyOo4S9eF9R6Xz9HJIo0YnSVmbfI7iBu1fdW5xuncaXy+yaNoejl1bd735LaFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQr1/1M//gDsendT 19:03:02,610 [LogDispatcher] [vTRACE] [request-handler.h:148] Decryption finished (b64): eNrtz7FOwzAQxvFXiTxBleA0UdSUmZWtTIjBOEcaNU6D7XRBvDsn9SUY/tvdZ/vOv/cfM7gseQpinvfdvu/q/tg3dV2X5muaNTT2LUlM1n1uYUsufNsXudkoaRo2qfy6Wg3XWZIdJDt9MlQ+uIvYFL116/rkvdFZ2+J11rTkIrhpedCiLPzZxWK3e9TzWW4y6w+aXutp0b1Ne9DyOo4S9eF9R6Xz9HJIo0YnSVmbfI7iBu1fdW5xuncaXy+yaNoejl1bd735LaFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQr1/1M//gDsendT 19:03:02,610 [LogDispatcher] [vTRACE] [request-handler.h:159] Decompressing: eNrtz7FOwzAQxvFXiTxBleA0UdSUmZWtTIjBOEcaNU6D7XRBvDsn9SUY/tvdZ/vOv/cfM7gseQpinvfdvu/q/tg3dV2X5muaNTT2LUlM1n1uYUsufNsXudkoaRo2qfy6Wg3XWZIdJDt9MlQ+uIvYFL116/rkvdFZ2+J11rTkIrhpedCiLPzZxWK3e9TzWW4y6w+aXutp0b1Ne9DyOo4S9eF9R6Xz9HJIo0YnSVmbfI7iBu1fdW5xuncaXy+yaNoejl1bd735LaFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQr1/1M//gDsendT 19:03:02,611 [LogDispatcher] [vTRACE] [request-handler.h:170] Decompression finished (raw): [{... (removed)}] 19:03:02,612 [LogDispatcher] [vDEBUG] [log-request-handler.cc:115] Is bulk? true 19:03:02,612 [LogDispatcher] [vDEBUG] [log-request-handler.cc:124] Request client: 0x10065cfa8 19:03:02,612 [LogHandler] [vTRACE] [session.cc:75] Received: 522 bytes 19:03:02,612 [LogDispatcher] [vDEBUG] [log-request-handler.cc:206] Force check: true, clientRef: 0x70000f9729a0, *clientRef: muflihun00102030, bypassChecks: false 19:03:02,612 [LogHandler] [vTRACE] [session.cc:102] Read bytes: 29277885c8e3830c20fb91fab830194b:muflihun00102030:zaAqRKANlz9SElDUXg6ePbeykddpAyJjxC3TEL/yutU5HIryqLpN1xmsEncDHizpniIeaCIY7kfUjuzTaXwYJC0dxHxlssq5Lm7PzgPCyyxa7POh0quW1zooTSlBMvUiuCWamAdfzz3WPPjjbMSyvqtgMmbFIhs4qMMMt3kKZbli8kk9Hmv5XV8Ltz+/tLX4ppiOn6qSFh/pIQ0LNaBVK0wHXLfGO+AN58vc1D50JT8vA00XUjkzNNmeasekIntZDoRut/WhgqVijoUsT27sY9WzrsdaLuIJIAUhtewqvAGgyeplI+lucgHWtWKVR+HQQ887giQwW6+80FrU0yIZ/yEhDnFbCul3Zc9L3vuuBHvrAgjYvDPkA4MZPwBkFWSh6anUbJ5zsNgEnMP2tdDI+ZLnSVfvQ+ZNEPeVXYE15la/XVPoe0/b/2MqY3HPUrYEtTJONqT/XrZKl1HXbqxp4w== [size: 522] 19:03:02,613 [LogDispatcher] [vDEBUG] [client.cc:65] Checking token 37953058 (client [muflihun00102030]) ```
1.0
Deadlock detected in develop - ``` 19:03:02,606 [LogHandler] [vTRACE] [session.cc:86] Adding bytes 19:03:02,608 [LogDispatcher] [vDEBUG] [log-request-handler.cc:88] Pausing schedule for client integrity 19:03:02,609 [LogHandler] [vTRACE] [session.cc:75] Received: 522 bytes 19:03:02,609 [LogHandler] [vTRACE] [session.cc:102] Read bytes: 0fe53a0a15ad3b19847a4e5ee30bdcda:muflihun00102030:sHbsBEN2KhHtBETNCpQDU3wHPw27YHBzztBj7yckU2sHE7lLyUw35wCpxM0qALezZcvXC51YAc75cm60NMY1+mJzr4gpHWlSduK4nj8Y87aPa3/GPoBOjpIl1bTYoUghZ6lCyNgBgNmw7I8Av6oxnaiWJ6AsZARX2Oy1K/DMv5kE33XjxEVkLwRSoCGOS71dSo6LtFzF16CzGXqr/bKHB3vBv+xb8Ap9rN2w36dXGk02ydyE5EF18Pt5pLVzEIyhROlG2ErWShBcWdigYy6pY7rl/qlY/9v1GbggQ0n4rUpPjKtS50H+Z4IQ8ZVT63FMEvb6FY6hs654hx7kdssWvA4jwa7r5x0OVr2+lme8qptSpxEJ15eesTYZvgQRXTImMrGXsqo3VXQZ/AQZ6ZSNaW4/99x17SP6SYiVxNIwQgv5WAnh7j+jAO5I7zTMY66kJQtrg7tozUAIBcM8oQed9Q== [size: 522] 19:03:02,609 [LogHandler] [vDEBUG] [session.cc:171] Sending {r:0} 19:03:02,609 [LogHandler] [vTRACE] [session.cc:86] Adding bytes 19:03:02,609 [LogDispatcher] [vCRAZY] [log-request-handler.cc:100] -----============= [ BEGIN ] =============----- 19:03:02,610 [LogDispatcher] [vDEBUG] [request-handler.h:144] Raw request: 938d714753441d7cdf1167023aa7f3f9:muflihun00102030:xLRbv36YW//gqY9gVrsCOBuwoE5NM+A8iJjoE7W1Jq1/PUXOvuCk0lVvlkmGZT2Hq0Vhj3vuzdSWxqiUH2Iu3I5ufo1EbWuIvQb04eZ0k6tZxJ1wRDjhdfIaCUab9Av0nlYohOWi56pD2suVynM4xucvXuFh6mm2T+qeTyF0i7LuZybJ5bouOHlxLzSCTiI+RQNPJqQo/TyCxim5Swa/PNZwD/S9/pPoPY42XDfZehan0r/kDP9kdQw2SPTaBxG79jyqF4g61o0fsnqtXtrVVVtxgbi5ySvSy40TpdYTbG7DME9hfRlGdxciyewftqijfaQePgJweYwTUYxVtl4uttiaPbELbe1OyC6qKvn7drfgFFtIttbUBa8lf8GV8FJ7vmpMr8CZMYk/8V4PYX9kRhiSKgCHK2vUTXlCja7CbYzoTgplL5iwrQf4LNJ/6FdckPS+ZtVHl9knQokhTyGbqw== 19:03:02,610 [LogDispatcher] [vDEBUG] [request-handler.cc:58] Client: muflihun00102030 19:03:02,610 [LogDispatcher] [vCRAZY] [request-handler.cc:59] IV: 938d714753441d7cdf1167023aa7f3f9 19:03:02,610 [LogDispatcher] [vCRAZY] [request-handler.cc:66] Data (base64): xLRbv36YW//gqY9gVrsCOBuwoE5NM+A8iJjoE7W1Jq1/PUXOvuCk0lVvlkmGZT2Hq0Vhj3vuzdSWxqiUH2Iu3I5ufo1EbWuIvQb04eZ0k6tZxJ1wRDjhdfIaCUab9Av0nlYohOWi56pD2suVynM4xucvXuFh6mm2T+qeTyF0i7LuZybJ5bouOHlxLzSCTiI+RQNPJqQo/TyCxim5Swa/PNZwD/S9/pPoPY42XDfZehan0r/kDP9kdQw2SPTaBxG79jyqF4g61o0fsnqtXtrVVVtxgbi5ySvSy40TpdYTbG7DME9hfRlGdxciyewftqijfaQePgJweYwTUYxVtl4uttiaPbELbe1OyC6qKvn7drfgFFtIttbUBa8lf8GV8FJ7vmpMr8CZMYk/8V4PYX9kRhiSKgCHK2vUTXlCja7CbYzoTgplL5iwrQf4LNJ/6FdckPS+ZtVHl9knQokhTyGbqw== 19:03:02,610 [LogDispatcher] [vCRAZY] [request-handler.cc:69] Ripe command: echo 938d714753441d7cdf1167023aa7f3f9:muflihun00102030:xLRbv36YW//gqY9gVrsCOBuwoE5NM+A8iJjoE7W1Jq1/PUXOvuCk0lVvlkmGZT2Hq0Vhj3vuzdSWxqiUH2Iu3I5ufo1EbWuIvQb04eZ0k6tZxJ1wRDjhdfIaCUab9Av0nlYohOWi56pD2suVynM4xucvXuFh6mm2T+qeTyF0i7LuZybJ5bouOHlxLzSCTiI+RQNPJqQo/TyCxim5Swa/PNZwD/S9/pPoPY42XDfZehan0r/kDP9kdQw2SPTaBxG79jyqF4g61o0fsnqtXtrVVVtxgbi5ySvSy40TpdYTbG7DME9hfRlGdxciyewftqijfaQePgJweYwTUYxVtl4uttiaPbELbe1OyC6qKvn7drfgFFtIttbUBa8lf8GV8FJ7vmpMr8CZMYk/8V4PYX9kRhiSKgCHK2vUTXlCja7CbYzoTgplL5iwrQf4LNJ/6FdckPS+ZtVHl9knQokhTyGbqw== | ripe -d --aes --key 00098C6A684633C0087D9C76B47D2A57 --base64 19:03:02,610 [LogDispatcher] [vCRAZY] [request-handler.cc:76] Plain request: eNrtz7FOwzAQxvFXiTxBleA0UdSUmZWtTIjBOEcaNU6D7XRBvDsn9SUY/tvdZ/vOv/cfM7gseQpinvfdvu/q/tg3dV2X5muaNTT2LUlM1n1uYUsufNsXudkoaRo2qfy6Wg3XWZIdJDt9MlQ+uIvYFL116/rkvdFZ2+J11rTkIrhpedCiLPzZxWK3e9TzWW4y6w+aXutp0b1Ne9DyOo4S9eF9R6Xz9HJIo0YnSVmbfI7iBu1fdW5xuncaXy+yaNoejl1bd735LaFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQr1/1M//gDsendT 19:03:02,610 [LogDispatcher] [vTRACE] [request-handler.h:148] Decryption finished (b64): eNrtz7FOwzAQxvFXiTxBleA0UdSUmZWtTIjBOEcaNU6D7XRBvDsn9SUY/tvdZ/vOv/cfM7gseQpinvfdvu/q/tg3dV2X5muaNTT2LUlM1n1uYUsufNsXudkoaRo2qfy6Wg3XWZIdJDt9MlQ+uIvYFL116/rkvdFZ2+J11rTkIrhpedCiLPzZxWK3e9TzWW4y6w+aXutp0b1Ne9DyOo4S9eF9R6Xz9HJIo0YnSVmbfI7iBu1fdW5xuncaXy+yaNoejl1bd735LaFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQr1/1M//gDsendT 19:03:02,610 [LogDispatcher] [vTRACE] [request-handler.h:159] Decompressing: eNrtz7FOwzAQxvFXiTxBleA0UdSUmZWtTIjBOEcaNU6D7XRBvDsn9SUY/tvdZ/vOv/cfM7gseQpinvfdvu/q/tg3dV2X5muaNTT2LUlM1n1uYUsufNsXudkoaRo2qfy6Wg3XWZIdJDt9MlQ+uIvYFL116/rkvdFZ2+J11rTkIrhpedCiLPzZxWK3e9TzWW4y6w+aXutp0b1Ne9DyOo4S9eF9R6Xz9HJIo0YnSVmbfI7iBu1fdW5xuncaXy+yaNoejl1bd735LaFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQoVKlSoUKFChQr1/1M//gDsendT 19:03:02,611 [LogDispatcher] [vTRACE] [request-handler.h:170] Decompression finished (raw): [{... (removed)}] 19:03:02,612 [LogDispatcher] [vDEBUG] [log-request-handler.cc:115] Is bulk? true 19:03:02,612 [LogDispatcher] [vDEBUG] [log-request-handler.cc:124] Request client: 0x10065cfa8 19:03:02,612 [LogHandler] [vTRACE] [session.cc:75] Received: 522 bytes 19:03:02,612 [LogDispatcher] [vDEBUG] [log-request-handler.cc:206] Force check: true, clientRef: 0x70000f9729a0, *clientRef: muflihun00102030, bypassChecks: false 19:03:02,612 [LogHandler] [vTRACE] [session.cc:102] Read bytes: 29277885c8e3830c20fb91fab830194b:muflihun00102030:zaAqRKANlz9SElDUXg6ePbeykddpAyJjxC3TEL/yutU5HIryqLpN1xmsEncDHizpniIeaCIY7kfUjuzTaXwYJC0dxHxlssq5Lm7PzgPCyyxa7POh0quW1zooTSlBMvUiuCWamAdfzz3WPPjjbMSyvqtgMmbFIhs4qMMMt3kKZbli8kk9Hmv5XV8Ltz+/tLX4ppiOn6qSFh/pIQ0LNaBVK0wHXLfGO+AN58vc1D50JT8vA00XUjkzNNmeasekIntZDoRut/WhgqVijoUsT27sY9WzrsdaLuIJIAUhtewqvAGgyeplI+lucgHWtWKVR+HQQ887giQwW6+80FrU0yIZ/yEhDnFbCul3Zc9L3vuuBHvrAgjYvDPkA4MZPwBkFWSh6anUbJ5zsNgEnMP2tdDI+ZLnSVfvQ+ZNEPeVXYE15la/XVPoe0/b/2MqY3HPUrYEtTJONqT/XrZKl1HXbqxp4w== [size: 522] 19:03:02,613 [LogDispatcher] [vDEBUG] [client.cc:65] Checking token 37953058 (client [muflihun00102030]) ```
process
deadlock detected in develop adding bytes pausing schedule for client integrity received bytes read bytes qly sending r adding bytes raw request rqnpjqqo pnzwd client iv data rqnpjqqo pnzwd ripe command echo rqnpjqqo pnzwd ripe d aes key plain request tvdz vov q gdsendt decryption finished tvdz vov q gdsendt decompressing tvdz vov q gdsendt decompression finished raw is bulk true request client received bytes force check true clientref clientref bypasschecks false read bytes lucghwtwkvr zlnsvfvq b checking token client
1
21,878
30,322,790,026
IssuesEvent
2023-07-10 20:39:04
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
Flaky Test - tailsamplingprocessor/TestConcurrentTraceArrival
bug help wanted good first issue priority:p3 processor/tailsampling flaky test never stale
``` race: limit on 8128 simultaneously alive goroutines is exceeded, dying FAIL github.com/open-telemetry/opentelemetry-collector-contrib/processor/tailsamplingprocessor 3.415s ``` Observed here: https://github.com/open-telemetry/opentelemetry-collector-contrib/runs/6555994409?check_suite_focus=true
1.0
Flaky Test - tailsamplingprocessor/TestConcurrentTraceArrival - ``` race: limit on 8128 simultaneously alive goroutines is exceeded, dying FAIL github.com/open-telemetry/opentelemetry-collector-contrib/processor/tailsamplingprocessor 3.415s ``` Observed here: https://github.com/open-telemetry/opentelemetry-collector-contrib/runs/6555994409?check_suite_focus=true
process
flaky test tailsamplingprocessor testconcurrenttracearrival race limit on simultaneously alive goroutines is exceeded dying fail github com open telemetry opentelemetry collector contrib processor tailsamplingprocessor observed here
1
2,293
5,114,591,185
IssuesEvent
2017-01-06 19:00:06
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
Multiple aggregations of the same type are missing correct `cols` metadata
Bug Priority/P1 Query Processor
Similar to #4002, but when two aggregations of the same type are specified, only the first has the correct `cols` metadata, in particular `base_type`, `special_type`, and `source` are incorrect. Example query: ```json { "source_table": 1, "aggregation": [["count"], ["count"]], "breakout": [["datetime_field", ["field-id", 1], "as", "day"]] } ``` Expected `data.cols[1]` and `data.cols[2]`: ```json { "description":null, "table_id":null, "special_type":"type/Number", "name":"count", "source":"aggregation", "extra_info":{}, "id":null, "target":null, "display_name":"count", "base_type":"type/Integer" }, { "description":null, "table_id":null, "special_type":"type/Number", "name":"count_2", "source":"aggregation", "extra_info":{}, "id":null, "target":null, "display_name":"count_2", "base_type":"type/Integer" } ``` Actual `data.cols[1]` and `data.cols[2]`: ```json { "description":null, "table_id":null, "special_type":"type/Number", "name":"count", "source":"aggregation", "extra_info":{}, "id":null, "target":null, "display_name":"count", "base_type":"type/Integer" }, { "description":null, "table_id":null, special_type:null, "name":"count_2", "extra_info":{}, "id":null, "target":null, "preview_display":true, "display_name":"count_2", base_type:"type/*" source: } ```
1.0
Multiple aggregations of the same type are missing correct `cols` metadata - Similar to #4002, but when two aggregations of the same type are specified, only the first has the correct `cols` metadata, in particular `base_type`, `special_type`, and `source` are incorrect. Example query: ```json { "source_table": 1, "aggregation": [["count"], ["count"]], "breakout": [["datetime_field", ["field-id", 1], "as", "day"]] } ``` Expected `data.cols[1]` and `data.cols[2]`: ```json { "description":null, "table_id":null, "special_type":"type/Number", "name":"count", "source":"aggregation", "extra_info":{}, "id":null, "target":null, "display_name":"count", "base_type":"type/Integer" }, { "description":null, "table_id":null, "special_type":"type/Number", "name":"count_2", "source":"aggregation", "extra_info":{}, "id":null, "target":null, "display_name":"count_2", "base_type":"type/Integer" } ``` Actual `data.cols[1]` and `data.cols[2]`: ```json { "description":null, "table_id":null, "special_type":"type/Number", "name":"count", "source":"aggregation", "extra_info":{}, "id":null, "target":null, "display_name":"count", "base_type":"type/Integer" }, { "description":null, "table_id":null, special_type:null, "name":"count_2", "extra_info":{}, "id":null, "target":null, "preview_display":true, "display_name":"count_2", base_type:"type/*" source: } ```
process
multiple aggregations of the same type are missing correct cols metadata similar to but when two aggregations of the same type are specified only the first has the correct cols metadata in particular base type special type and source are incorrect example query json source table aggregation breakout as day expected data cols and data cols json description null table id null special type type number name count source aggregation extra info id null target null display name count base type type integer description null table id null special type type number name count source aggregation extra info id null target null display name count base type type integer actual data cols and data cols json description null table id null special type type number name count source aggregation extra info id null target null display name count base type type integer description null table id null special type null name count extra info id null target null preview display true display name count base type type source
1
4,130
7,086,165,366
IssuesEvent
2018-01-11 13:43:46
rogerthat-platform/rogerthat-android-client
https://api.github.com/repos/rogerthat-platform/rogerthat-android-client
closed
Apps with a custom home screen
priority_critical process_duplicate type_feature
- [ ] Use homescreen_style "branding" in `build.yaml` - [ ] After registration, show progress bar until main service, all js embeddings and home branding are available. - [ ] Implement api to read news items from inside a branding - Security: return all news items to the main service. Other services should only receive their news items. - [ ] Triggering of badge number callbacks
1.0
Apps with a custom home screen - - [ ] Use homescreen_style "branding" in `build.yaml` - [ ] After registration, show progress bar until main service, all js embeddings and home branding are available. - [ ] Implement api to read news items from inside a branding - Security: return all news items to the main service. Other services should only receive their news items. - [ ] Triggering of badge number callbacks
process
apps with a custom home screen use homescreen style branding in build yaml after registration show progress bar until main service all js embeddings and home branding are available implement api to read news items from inside a branding security return all news items to the main service other services should only receive their news items triggering of badge number callbacks
1
17,602
23,425,325,115
IssuesEvent
2022-08-14 09:55:08
Battle-s/battle-school-backend
https://api.github.com/repos/Battle-s/battle-school-backend
opened
[FEAT] 시즌 생성 및 조회
feature :computer: processing :hourglass_flowing_sand:
## 설명 > 시즌 생성 및 조회 ## 체크사항 - [ ] 시즌 엔티티 및 repo 생성 - [ ] 시즌 서비스 - crud ## 참고자료 ## 관련 논의
1.0
[FEAT] 시즌 생성 및 조회 - ## 설명 > 시즌 생성 및 조회 ## 체크사항 - [ ] 시즌 엔티티 및 repo 생성 - [ ] 시즌 서비스 - crud ## 참고자료 ## 관련 논의
process
시즌 생성 및 조회 설명 시즌 생성 및 조회 체크사항 시즌 엔티티 및 repo 생성 시즌 서비스 crud 참고자료 관련 논의
1
78,465
27,533,147,127
IssuesEvent
2023-03-07 00:13:08
openzfs/zfs
https://api.github.com/repos/openzfs/zfs
closed
Performance degradation in NVMe L2ARC configured ZFS upgrading from 2.1.4 to 2.1.6
Type: Defect
<!-- Please fill out the following template, which will help other contributors address your issue. --> <!-- Thank you for reporting an issue. *IMPORTANT* - Please check our issue tracker before opening a new issue. Additional valuable information can be found in the OpenZFS documentation and mailing list archives. Please fill in as much of the template as possible. --> ### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | Amazon Linux 2 Distribution Version | Amazon Linux 2 AMI 2.0.20230119.1 x86_64 Kernel Version | 5.10.165-143.735.amzn2.x86_64 Architecture | x86_64 OpenZFS Version | 2.1.6 <!-- Command to find OpenZFS version: zfs version Commands to find kernel version: uname -r # Linux freebsd-version -r # FreeBSD --> ### Describe the problem you're observing Upgrading from ZFS 2.1.4 to ZFS 2.1.6, I'm seeing increased read latency and reduced bandwidth, IOPS. I bisected it to the following commit: [39e038cdc61a9af2549fa3ed1f42c16dbcbe4ca6](https://github.com/openzfs/zfs/pull/12285/commits/39e038cdc61a9af2549fa3ed1f42c16dbcbe4ca6) I have the l2arc_exclude_special tunable set to 0 as is the default. Reverting this commit seems to fix the issue. The issue is only affecting performance for a NVMe L2ARC configured ZFS. ZFS without L2ARC is unaffected.
1.0
Performance degradation in NVMe L2ARC configured ZFS upgrading from 2.1.4 to 2.1.6 - <!-- Please fill out the following template, which will help other contributors address your issue. --> <!-- Thank you for reporting an issue. *IMPORTANT* - Please check our issue tracker before opening a new issue. Additional valuable information can be found in the OpenZFS documentation and mailing list archives. Please fill in as much of the template as possible. --> ### System information <!-- add version after "|" character --> Type | Version/Name --- | --- Distribution Name | Amazon Linux 2 Distribution Version | Amazon Linux 2 AMI 2.0.20230119.1 x86_64 Kernel Version | 5.10.165-143.735.amzn2.x86_64 Architecture | x86_64 OpenZFS Version | 2.1.6 <!-- Command to find OpenZFS version: zfs version Commands to find kernel version: uname -r # Linux freebsd-version -r # FreeBSD --> ### Describe the problem you're observing Upgrading from ZFS 2.1.4 to ZFS 2.1.6, I'm seeing increased read latency and reduced bandwidth, IOPS. I bisected it to the following commit: [39e038cdc61a9af2549fa3ed1f42c16dbcbe4ca6](https://github.com/openzfs/zfs/pull/12285/commits/39e038cdc61a9af2549fa3ed1f42c16dbcbe4ca6) I have the l2arc_exclude_special tunable set to 0 as is the default. Reverting this commit seems to fix the issue. The issue is only affecting performance for a NVMe L2ARC configured ZFS. ZFS without L2ARC is unaffected.
non_process
performance degradation in nvme configured zfs upgrading from to thank you for reporting an issue important please check our issue tracker before opening a new issue additional valuable information can be found in the openzfs documentation and mailing list archives please fill in as much of the template as possible system information type version name distribution name amazon linux distribution version amazon linux ami kernel version architecture openzfs version command to find openzfs version zfs version commands to find kernel version uname r linux freebsd version r freebsd describe the problem you re observing upgrading from zfs to zfs i m seeing increased read latency and reduced bandwidth iops i bisected it to the following commit i have the exclude special tunable set to as is the default reverting this commit seems to fix the issue the issue is only affecting performance for a nvme configured zfs zfs without is unaffected
0
63,695
8,691,141,700
IssuesEvent
2018-12-04 00:01:29
CuBoulder/express
https://api.github.com/repos/CuBoulder/express
closed
As a SO/CE I would like to place interactive blocks on my website
Epic Still Open at 3.0 evaluate-1:Director Approval evaluate-1:More Information evaluate-2:Accessibility evaluate-2:Documentation Needs evaluate-2:Support Concerns evaluate-2:Usability
## Context Good afternoon, Do we have the feature in Drupal that allows us to hover a box and for text to appear like so: http://alumni.umich.edu/ Thank you, Janeyl ![interactive blocks](https://user-images.githubusercontent.com/16137919/28079330-2165aca8-6625-11e7-8770-455dc6821c6e.png)
1.0
As a SO/CE I would like to place interactive blocks on my website - ## Context Good afternoon, Do we have the feature in Drupal that allows us to hover a box and for text to appear like so: http://alumni.umich.edu/ Thank you, Janeyl ![interactive blocks](https://user-images.githubusercontent.com/16137919/28079330-2165aca8-6625-11e7-8770-455dc6821c6e.png)
non_process
as a so ce i would like to place interactive blocks on my website context good afternoon do we have the feature in drupal that allows us to hover a box and for text to appear like so thank you janeyl
0
21,660
30,107,792,392
IssuesEvent
2023-06-30 04:04:12
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
Rename websocketprocessor to RemoteObserverProcessor
enhancement needs triage processor/websocket
### Component(s) processor/websocket ### Is your feature request related to a problem? Please describe. As discussed in #19633, the websocketprocessor name is too broad and confusing, and doesn't characterize the intent of use to the satisfaction of the community. ### Describe the solution you'd like Rename the websocketprocessor to RemoteObserverProcessor so its use becomes clear. ### Describe alternatives you've considered _No response_ ### Additional context _No response_
1.0
Rename websocketprocessor to RemoteObserverProcessor - ### Component(s) processor/websocket ### Is your feature request related to a problem? Please describe. As discussed in #19633, the websocketprocessor name is too broad and confusing, and doesn't characterize the intent of use to the satisfaction of the community. ### Describe the solution you'd like Rename the websocketprocessor to RemoteObserverProcessor so its use becomes clear. ### Describe alternatives you've considered _No response_ ### Additional context _No response_
process
rename websocketprocessor to remoteobserverprocessor component s processor websocket is your feature request related to a problem please describe as discussed in the websocketprocessor name is too broad and confusing and doesn t characterize the intent of use to the satisfaction of the community describe the solution you d like rename the websocketprocessor to remoteobserverprocessor so its use becomes clear describe alternatives you ve considered no response additional context no response
1
6,356
9,415,072,604
IssuesEvent
2019-04-10 11:46:00
AmpersandTarski/Ampersand
https://api.github.com/repos/AmpersandTarski/Ampersand
opened
stack install fails on MacBook
OSX priority:normal software process
#### Version of ampersand that was used This problem occurred on commit c0239e76195cbf7692b5bb7e15dbcdb7aafd60cf, on the development branch of the github Ampersand repository. #### What I expected Since the purpose of stack is to build in platform independent ways, I expected "stack install" to build Ampersand for me. #### What happened instead The command "stack install" fails quite quickly: <img width="682" alt="Screenshot 2019-04-10 at 13 30 52" src="https://user-images.githubusercontent.com/275768/55875911-8bc9f000-5b96-11e9-9995-c06879796b43.png"> Apparently, it uses an executable meant for Windows. #### Steps to reproduce 1. 2. 3. 4. #### Screenshot / Video #### Context / Source of ampersand script <!-- Optional: share your script if possible. It helps us reproduce the problem. Please try to keep the scripts tiny We'd also love to know how you found the bug: #dogfooding, #manual-testing, #automated-testing, or #user-report if applicable. If requesting a new feature, explain why you'd like to see it added. -->
1.0
stack install fails on MacBook - #### Version of ampersand that was used This problem occurred on commit c0239e76195cbf7692b5bb7e15dbcdb7aafd60cf, on the development branch of the github Ampersand repository. #### What I expected Since the purpose of stack is to build in platform independent ways, I expected "stack install" to build Ampersand for me. #### What happened instead The command "stack install" fails quite quickly: <img width="682" alt="Screenshot 2019-04-10 at 13 30 52" src="https://user-images.githubusercontent.com/275768/55875911-8bc9f000-5b96-11e9-9995-c06879796b43.png"> Apparently, it uses an executable meant for Windows. #### Steps to reproduce 1. 2. 3. 4. #### Screenshot / Video #### Context / Source of ampersand script <!-- Optional: share your script if possible. It helps us reproduce the problem. Please try to keep the scripts tiny We'd also love to know how you found the bug: #dogfooding, #manual-testing, #automated-testing, or #user-report if applicable. If requesting a new feature, explain why you'd like to see it added. -->
process
stack install fails on macbook version of ampersand that was used this problem occurred on commit on the development branch of the github ampersand repository what i expected since the purpose of stack is to build in platform independent ways i expected stack install to build ampersand for me what happened instead the command stack install fails quite quickly img width alt screenshot at src apparently it uses an executable meant for windows steps to reproduce screenshot video context source of ampersand script optional share your script if possible it helps us reproduce the problem please try to keep the scripts tiny we d also love to know how you found the bug dogfooding manual testing automated testing or user report if applicable if requesting a new feature explain why you d like to see it added
1
21,727
30,235,103,069
IssuesEvent
2023-07-06 09:40:13
mrdoob/three.js
https://api.github.com/repos/mrdoob/three.js
closed
MaskPass breaks background update for transmissive objects
Bug Post-processing
Currently there are the following [lines of code in the MaskPass class](https://github.com/mrdoob/three.js/blob/42c59b4d516fd204f5f760e1f7daefbbbf18c49c/examples/jsm/postprocessing/MaskPass.js#L26) ```js state.buffers.color.setMask( false ); state.buffers.depth.setMask( false ); ``` Both values are set to `false`, but these operations aren't reverted at the end of the pass. As the result, if after the scene is rendered you remove the scene background, it's not removed for transmissive objects. What's more, if there are non-transmissive parts in a transmissive model and it's being rotated, then the non-transmissive parts leave a trail on the background. 1. A box mask is applied to a scene with some background and a transmissive object. Note that there is a small non-transmissive box in the model. ![Screenshot_1](https://user-images.githubusercontent.com/96049300/204225293-18eda9b1-f6a6-4295-a5d3-84fc7d211d06.png) 2. The scene background is removed, but it's still there for the transmissive object. ![Screenshot_2](https://user-images.githubusercontent.com/96049300/204225435-3bc1b697-9a0f-4533-aab3-d5fa4adc5db4.png) 3. When the object is rotated, the small opaque black box leaves a trail on the background. ![Screenshot_3](https://user-images.githubusercontent.com/96049300/204225774-dea1c906-3154-452e-87ce-59f5827e52b0.png) **Some code** ```js function createBox(scale = 1, x = 0, y = 0, z = 0) { const geometry = new BoxGeometry(scale , scale , scale); const mesh = new Mesh(geometry); mesh.material = new MeshBasicMaterial({ color: 0x000000 }); mesh.position.x = x; mesh.position.y = y; mesh.position.z = z; return mesh; } const maskScene = new Scene(); maskScene.add(createBox(3)); // load a model and add an opaque box to it gltf.scene.add(createBox(0.1, -0.5, 0.5, 0)); scene.add(gltf.scene); const maskRenderTarget = new WebGLRenderTarget(500, 500, { stencilBuffer: true, samples: 8, }); const composer = new EffectComposer(renderer, maskRenderTarget); composer.setPixelRatio(window.devicePixelRatio); // passes composer.addPass(new ClearPass()); composer.addPass(new MaskPass(maskScene, camera)); const renderPass = new RenderPass(scene, camera); renderPass.clear = false; composer.addPass(renderPass); composer.addPass(new ClearMaskPass()); composer.addPass(new ShaderPass(GammaCorrectionShader)); ``` **Probable solution** If the following code ```js state.buffers.color.setMask( true ); state.buffers.depth.setMask( true ); ``` is executed before `RenderPass` (e.g. at the end of `MaskPass`), then the problem disappears. For this very issue only the first line is enough.
1.0
MaskPass breaks background update for transmissive objects - Currently there are the following [lines of code in the MaskPass class](https://github.com/mrdoob/three.js/blob/42c59b4d516fd204f5f760e1f7daefbbbf18c49c/examples/jsm/postprocessing/MaskPass.js#L26) ```js state.buffers.color.setMask( false ); state.buffers.depth.setMask( false ); ``` Both values are set to `false`, but these operations aren't reverted at the end of the pass. As the result, if after the scene is rendered you remove the scene background, it's not removed for transmissive objects. What's more, if there are non-transmissive parts in a transmissive model and it's being rotated, then the non-transmissive parts leave a trail on the background. 1. A box mask is applied to a scene with some background and a transmissive object. Note that there is a small non-transmissive box in the model. ![Screenshot_1](https://user-images.githubusercontent.com/96049300/204225293-18eda9b1-f6a6-4295-a5d3-84fc7d211d06.png) 2. The scene background is removed, but it's still there for the transmissive object. ![Screenshot_2](https://user-images.githubusercontent.com/96049300/204225435-3bc1b697-9a0f-4533-aab3-d5fa4adc5db4.png) 3. When the object is rotated, the small opaque black box leaves a trail on the background. ![Screenshot_3](https://user-images.githubusercontent.com/96049300/204225774-dea1c906-3154-452e-87ce-59f5827e52b0.png) **Some code** ```js function createBox(scale = 1, x = 0, y = 0, z = 0) { const geometry = new BoxGeometry(scale , scale , scale); const mesh = new Mesh(geometry); mesh.material = new MeshBasicMaterial({ color: 0x000000 }); mesh.position.x = x; mesh.position.y = y; mesh.position.z = z; return mesh; } const maskScene = new Scene(); maskScene.add(createBox(3)); // load a model and add an opaque box to it gltf.scene.add(createBox(0.1, -0.5, 0.5, 0)); scene.add(gltf.scene); const maskRenderTarget = new WebGLRenderTarget(500, 500, { stencilBuffer: true, samples: 8, }); const composer = new EffectComposer(renderer, maskRenderTarget); composer.setPixelRatio(window.devicePixelRatio); // passes composer.addPass(new ClearPass()); composer.addPass(new MaskPass(maskScene, camera)); const renderPass = new RenderPass(scene, camera); renderPass.clear = false; composer.addPass(renderPass); composer.addPass(new ClearMaskPass()); composer.addPass(new ShaderPass(GammaCorrectionShader)); ``` **Probable solution** If the following code ```js state.buffers.color.setMask( true ); state.buffers.depth.setMask( true ); ``` is executed before `RenderPass` (e.g. at the end of `MaskPass`), then the problem disappears. For this very issue only the first line is enough.
process
maskpass breaks background update for transmissive objects currently there are the following js state buffers color setmask false state buffers depth setmask false both values are set to false but these operations aren t reverted at the end of the pass as the result if after the scene is rendered you remove the scene background it s not removed for transmissive objects what s more if there are non transmissive parts in a transmissive model and it s being rotated then the non transmissive parts leave a trail on the background a box mask is applied to a scene with some background and a transmissive object note that there is a small non transmissive box in the model the scene background is removed but it s still there for the transmissive object when the object is rotated the small opaque black box leaves a trail on the background some code js function createbox scale x y z const geometry new boxgeometry scale scale scale const mesh new mesh geometry mesh material new meshbasicmaterial color mesh position x x mesh position y y mesh position z z return mesh const maskscene new scene maskscene add createbox load a model and add an opaque box to it gltf scene add createbox scene add gltf scene const maskrendertarget new webglrendertarget stencilbuffer true samples const composer new effectcomposer renderer maskrendertarget composer setpixelratio window devicepixelratio passes composer addpass new clearpass composer addpass new maskpass maskscene camera const renderpass new renderpass scene camera renderpass clear false composer addpass renderpass composer addpass new clearmaskpass composer addpass new shaderpass gammacorrectionshader probable solution if the following code js state buffers color setmask true state buffers depth setmask true is executed before renderpass e g at the end of maskpass then the problem disappears for this very issue only the first line is enough
1
117,851
25,203,539,006
IssuesEvent
2022-11-13 12:05:44
leo-editor/leo-editor
https://api.github.com/repos/leo-editor/leo-editor
closed
Fix problems with ftm.set_widgets and sftm.init_widgets
Code
From an email from Félix. There are several problems with FindtabManager.set_widgets_from_dict and StringFindTabManager.init_widgets: - sftm.init_widgets does not handle the "File Only" option. - Neither method defines "val" after the comment `# Ensure one radio button is set.` - Both methods should fall back to "Entire Outline" instead of "File Only"!
1.0
Fix problems with ftm.set_widgets and sftm.init_widgets - From an email from Félix. There are several problems with FindtabManager.set_widgets_from_dict and StringFindTabManager.init_widgets: - sftm.init_widgets does not handle the "File Only" option. - Neither method defines "val" after the comment `# Ensure one radio button is set.` - Both methods should fall back to "Entire Outline" instead of "File Only"!
non_process
fix problems with ftm set widgets and sftm init widgets from an email from félix there are several problems with findtabmanager set widgets from dict and stringfindtabmanager init widgets sftm init widgets does not handle the file only option neither method defines val after the comment ensure one radio button is set both methods should fall back to entire outline instead of file only
0
323,582
27,738,234,017
IssuesEvent
2023-03-15 12:44:37
Jersey-Electricity-plc/intsys-issues
https://api.github.com/repos/Jersey-Electricity-plc/intsys-issues
closed
User Setup for NTS_D1\SLEG message when trying to create a quote
bug readyToTest
Trying to create a quote for HD0018781 but the bellow message appears. I have looked at user set up but my user ID is different to everyone else. Mine is SLEG not NTS_D1\SLEG, apologies if I'm missing something. ![image](https://user-images.githubusercontent.com/122997604/222700958-c68c11ad-8f73-4cb8-8076-1ea99ff7ae5a.png)
1.0
User Setup for NTS_D1\SLEG message when trying to create a quote - Trying to create a quote for HD0018781 but the bellow message appears. I have looked at user set up but my user ID is different to everyone else. Mine is SLEG not NTS_D1\SLEG, apologies if I'm missing something. ![image](https://user-images.githubusercontent.com/122997604/222700958-c68c11ad-8f73-4cb8-8076-1ea99ff7ae5a.png)
non_process
user setup for nts sleg message when trying to create a quote trying to create a quote for but the bellow message appears i have looked at user set up but my user id is different to everyone else mine is sleg not nts sleg apologies if i m missing something
0
12,038
14,738,709,423
IssuesEvent
2021-01-07 05:31:09
kdjstudios/SABillingGitlab
https://api.github.com/repos/kdjstudios/SABillingGitlab
closed
Spill Center Language Line
anc-ops anc-process anp-1.5 ant-support
In GitLab by @kdjstudios on Jul 11, 2018, 14:00 **Submitted by:** "Kimberly Gagner" <kimberly.gagner@answernet.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-07-11-14887/conversation **Server:** Internal **Client/Site:** 123 **Account:** Multiple **Issue:** I am reaching out to you to ask for your help in fixing the $50 fees that are appearing on some of the Spill Center language accounts. I know why this has happened and I will explain. When the language accounts first were set up in SAB they were set up as individual accounts. Now, the language accounts are all linked to E0840 and billed from that account. Since the link was added and the $50.00 charges were already showing on the invoice before the accounts were linked is why I cannot remove the charges. I have tried entering staged fees individually in each one and it is not working correctly. I need someone with the ability to go into the following accounts and remove those $50.00 charges. Here is the list: E0844 – Spill Center Cantonese E0842 – Spill Center Czech E0852 – Spill Center Flemish E0846 – Spill Center Japanese E0851 – Spill Center Korean E0853 – Spill Center Polish E0854 – Spill Center – Portuguese E0849 – Spill Center – Turkish Another reason this needs to be corrected is it is appearing on my aging report. Kyle, I attached a screenshot as well. If you have questions, please call me at x12301.
1.0
Spill Center Language Line - In GitLab by @kdjstudios on Jul 11, 2018, 14:00 **Submitted by:** "Kimberly Gagner" <kimberly.gagner@answernet.com> **Helpdesk:** http://www.servicedesk.answernet.com/profiles/ticket/2018-07-11-14887/conversation **Server:** Internal **Client/Site:** 123 **Account:** Multiple **Issue:** I am reaching out to you to ask for your help in fixing the $50 fees that are appearing on some of the Spill Center language accounts. I know why this has happened and I will explain. When the language accounts first were set up in SAB they were set up as individual accounts. Now, the language accounts are all linked to E0840 and billed from that account. Since the link was added and the $50.00 charges were already showing on the invoice before the accounts were linked is why I cannot remove the charges. I have tried entering staged fees individually in each one and it is not working correctly. I need someone with the ability to go into the following accounts and remove those $50.00 charges. Here is the list: E0844 – Spill Center Cantonese E0842 – Spill Center Czech E0852 – Spill Center Flemish E0846 – Spill Center Japanese E0851 – Spill Center Korean E0853 – Spill Center Polish E0854 – Spill Center – Portuguese E0849 – Spill Center – Turkish Another reason this needs to be corrected is it is appearing on my aging report. Kyle, I attached a screenshot as well. If you have questions, please call me at x12301.
process
spill center language line in gitlab by kdjstudios on jul submitted by kimberly gagner helpdesk server internal client site account multiple issue i am reaching out to you to ask for your help in fixing the fees that are appearing on some of the spill center language accounts i know why this has happened and i will explain when the language accounts first were set up in sab they were set up as individual accounts now the language accounts are all linked to and billed from that account since the link was added and the charges were already showing on the invoice before the accounts were linked is why i cannot remove the charges i have tried entering staged fees individually in each one and it is not working correctly i need someone with the ability to go into the following accounts and remove those charges here is the list – spill center cantonese – spill center czech – spill center flemish – spill center japanese – spill center korean – spill center polish – spill center – portuguese – spill center – turkish another reason this needs to be corrected is it is appearing on my aging report kyle i attached a screenshot as well if you have questions please call me at
1
8,438
11,599,117,911
IssuesEvent
2020-02-25 01:10:38
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
-ThrottleLimit
Pri2 automation/svc cxp doc-enhancement process-automation/subsvc triaged
It's not clear what -ThrottleLimit object value means. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 805a6236-70b7-7dd5-ac86-eea6efceff3a * Version Independent ID: e8078e34-bdf0-32b1-fac4-550091f2a06a * Content: [Learning PowerShell Workflow for Azure Automation](https://docs.microsoft.com/en-us/azure/automation/automation-powershell-workflow#feedback) * Content Source: [articles/automation/automation-powershell-workflow.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-powershell-workflow.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @MGoedtel * Microsoft Alias: **magoedte**
1.0
-ThrottleLimit - It's not clear what -ThrottleLimit object value means. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 805a6236-70b7-7dd5-ac86-eea6efceff3a * Version Independent ID: e8078e34-bdf0-32b1-fac4-550091f2a06a * Content: [Learning PowerShell Workflow for Azure Automation](https://docs.microsoft.com/en-us/azure/automation/automation-powershell-workflow#feedback) * Content Source: [articles/automation/automation-powershell-workflow.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-powershell-workflow.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @MGoedtel * Microsoft Alias: **magoedte**
process
throttlelimit it s not clear what throttlelimit object value means document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login mgoedtel microsoft alias magoedte
1
6,636
9,745,796,601
IssuesEvent
2019-06-03 10:30:34
EthVM/EthVM
https://api.github.com/repos/EthVM/EthVM
opened
Parametrize window period on Kafka Processing
bug enhancement priority:high project:processing
* **I'm submitting a ...** - [X] feature request * **Feature Request** - In processing MainNet we obtained the following exception: ``` [2019-06-03 07:02:48,998] WARN WorkerSinkTask{id=postgres-block-sink-0} Commit of offsets timed out (org.apache.kafka.connect.runtime.WorkerSinkTask) ``` Which means that if for some reason we wait too long to obtain new data, the windows are expired. We need to parametrize that in our Kafka processing with an Env variable (so different networks can have different settings). I'll mark this as an enhancement and as a bug, as we need this change for processing properly MainNet.
1.0
Parametrize window period on Kafka Processing - * **I'm submitting a ...** - [X] feature request * **Feature Request** - In processing MainNet we obtained the following exception: ``` [2019-06-03 07:02:48,998] WARN WorkerSinkTask{id=postgres-block-sink-0} Commit of offsets timed out (org.apache.kafka.connect.runtime.WorkerSinkTask) ``` Which means that if for some reason we wait too long to obtain new data, the windows are expired. We need to parametrize that in our Kafka processing with an Env variable (so different networks can have different settings). I'll mark this as an enhancement and as a bug, as we need this change for processing properly MainNet.
process
parametrize window period on kafka processing i m submitting a feature request feature request in processing mainnet we obtained the following exception warn workersinktask id postgres block sink commit of offsets timed out org apache kafka connect runtime workersinktask which means that if for some reason we wait too long to obtain new data the windows are expired we need to parametrize that in our kafka processing with an env variable so different networks can have different settings i ll mark this as an enhancement and as a bug as we need this change for processing properly mainnet
1
10,291
2,939,353,831
IssuesEvent
2015-07-01 15:52:04
WikiWatershed/model-my-watershed
https://api.github.com/repos/WikiWatershed/model-my-watershed
opened
iPad Issues
design
Taken on iPad Mini 2 running iOS 8.3. * The modifications dropdowns has width issues: ![2015-07-01 11 07 36](https://cloud.githubusercontent.com/assets/1430060/8458809/1e7cfde2-1fe7-11e5-87bb-5e6ded9f31a3.png) * In some cases, the modification thumbnails are offscreen and the user cannot scroll below to see them: ![2015-07-01 11 07 23](https://cloud.githubusercontent.com/assets/1430060/8458828/3d1743ac-1fe7-11e5-864b-38de5bb48f7f.png) * Unlike behavior on Android, on an iPad tapping once simulates a Hover, and tapping twice simulates a Click (selection). Thus the user must tap twice on any modification to select it, which is not very obvious. Also, when the user is in the "Draw" mode, there is no visual feedback for this (unlike on the desktop where the cursor changes and has an attached tooltip).
1.0
iPad Issues - Taken on iPad Mini 2 running iOS 8.3. * The modifications dropdowns has width issues: ![2015-07-01 11 07 36](https://cloud.githubusercontent.com/assets/1430060/8458809/1e7cfde2-1fe7-11e5-87bb-5e6ded9f31a3.png) * In some cases, the modification thumbnails are offscreen and the user cannot scroll below to see them: ![2015-07-01 11 07 23](https://cloud.githubusercontent.com/assets/1430060/8458828/3d1743ac-1fe7-11e5-864b-38de5bb48f7f.png) * Unlike behavior on Android, on an iPad tapping once simulates a Hover, and tapping twice simulates a Click (selection). Thus the user must tap twice on any modification to select it, which is not very obvious. Also, when the user is in the "Draw" mode, there is no visual feedback for this (unlike on the desktop where the cursor changes and has an attached tooltip).
non_process
ipad issues taken on ipad mini running ios the modifications dropdowns has width issues in some cases the modification thumbnails are offscreen and the user cannot scroll below to see them unlike behavior on android on an ipad tapping once simulates a hover and tapping twice simulates a click selection thus the user must tap twice on any modification to select it which is not very obvious also when the user is in the draw mode there is no visual feedback for this unlike on the desktop where the cursor changes and has an attached tooltip
0
166,375
26,345,623,824
IssuesEvent
2023-01-10 21:46:41
gotogether-s/gotogether-s
https://api.github.com/repos/gotogether-s/gotogether-s
closed
SignUp
design
## About Update style of signup page ## To do list - [x] Sort import lines of signup page - [x] Organize code lines of signup page - [x] Reduce paddingBottom from 1rem to 0.5rem in signin page - [x] Reduce paddingTop value of error message in signin page - [x] Reduce marginBottom value of input wrapper box - [x] Remove marginTop value of signup link - [x] Reduce default marginBottom value to 2rem - [x] Add style to Button on hover - [x] Replace HTML with MUI, Update style and Delete SignUp.scss - [x] Reduce size of icon and Increase fontSize of page title
1.0
SignUp - ## About Update style of signup page ## To do list - [x] Sort import lines of signup page - [x] Organize code lines of signup page - [x] Reduce paddingBottom from 1rem to 0.5rem in signin page - [x] Reduce paddingTop value of error message in signin page - [x] Reduce marginBottom value of input wrapper box - [x] Remove marginTop value of signup link - [x] Reduce default marginBottom value to 2rem - [x] Add style to Button on hover - [x] Replace HTML with MUI, Update style and Delete SignUp.scss - [x] Reduce size of icon and Increase fontSize of page title
non_process
signup about update style of signup page to do list sort import lines of signup page organize code lines of signup page reduce paddingbottom from to in signin page reduce paddingtop value of error message in signin page reduce marginbottom value of input wrapper box remove margintop value of signup link reduce default marginbottom value to add style to button on hover replace html with mui update style and delete signup scss reduce size of icon and increase fontsize of page title
0
818,783
30,704,318,196
IssuesEvent
2023-07-27 04:09:26
ballerina-platform/ballerina-lang
https://api.github.com/repos/ballerina-platform/ballerina-lang
closed
[Bug]: NPE in ComputedResourceAccessSegmentNodeContext
Type/Bug Priority/High Team/LanguageServer Points/1 Area/Completion userCategory/Editor
### Description In the following scenario, when we type something after `"path" a`, a null pointer exception is thrown. https://github.com/ballerina-platform/ballerina-lang/assets/61020198/3b61c5ad-b901-46d2-83f5-d6370c7af104 ### Steps to Reproduce ``` client class ClientClass { resource function accessor ["path" a] () { } } public function main() { var cl = new ClientClass(); cl ->/["path"].accessor; } ``` ### Affected Version(s) _No response_ ### OS, DB, other environment details and versions _No response_ ### Related area -> Editor ### Related issue(s) (optional) _No response_ ### Suggested label(s) (optional) _No response_ ### Suggested assignee(s) (optional) _No response_
1.0
[Bug]: NPE in ComputedResourceAccessSegmentNodeContext - ### Description In the following scenario, when we type something after `"path" a`, a null pointer exception is thrown. https://github.com/ballerina-platform/ballerina-lang/assets/61020198/3b61c5ad-b901-46d2-83f5-d6370c7af104 ### Steps to Reproduce ``` client class ClientClass { resource function accessor ["path" a] () { } } public function main() { var cl = new ClientClass(); cl ->/["path"].accessor; } ``` ### Affected Version(s) _No response_ ### OS, DB, other environment details and versions _No response_ ### Related area -> Editor ### Related issue(s) (optional) _No response_ ### Suggested label(s) (optional) _No response_ ### Suggested assignee(s) (optional) _No response_
non_process
npe in computedresourceaccesssegmentnodecontext description in the following scenario when we type something after path a a null pointer exception is thrown steps to reproduce client class clientclass resource function accessor public function main var cl new clientclass cl accessor affected version s no response os db other environment details and versions no response related area editor related issue s optional no response suggested label s optional no response suggested assignee s optional no response
0
2,915
5,909,019,253
IssuesEvent
2017-05-19 22:12:06
metabase/metabase
https://api.github.com/repos/metabase/metabase
opened
Second date breakout doesn't have `unit` in the result `col`
Query Processor
If I breakout by 2 date columns only the first has a `unit` in the response `cols`. This causes headers in pivot tables to not be formatted correctly, and probably various other issues. ```json { "data": { "cols" : [ { "source" : "breakout", "schema_name" : "PUBLIC", "extra_info" : {}, "special_type" : null, "name" : "CREATED_AT", "fk_field_id" : null, "visibility_type" : "normal", "display_name" : "Created At", "target" : null, "base_type" : "type/DateTime", "id" : 1, "description" : "The date and time an order was submitted.", "table_id" : 1, "unit" : "month" }, { "preview_display" : true, "display_name" : "CREATED_AT_2", "target" : null, "id" : null, "base_type" : "type/DateTime", "table_id" : null, "description" : null, "source" : "breakout", "extra_info" : {}, "special_type" : null, "name" : "CREATED_AT_2" }, { "extra_info" : {}, "special_type" : "type/Number", "name" : "count", "source" : "aggregation", "base_type" : "type/Integer", "id" : null, "description" : null, "table_id" : null, "target" : null, "display_name" : "count" } ] }, "json_query" : { "query" : { "source_table" : 1, "aggregation" : [ [ "count" ] ], "breakout" : [ [ "datetime-field", [ "field-id", 1 ], "as", "month" ], [ "datetime-field", [ "fk->", 3, 22 ], "as", "month" ] ] }, "type" : "query", "parameters" : [], "database" : 1, "constraints" : { "max-results-bare-rows" : 2000, "max-results" : 10000 } } } ```
1.0
Second date breakout doesn't have `unit` in the result `col` - If I breakout by 2 date columns only the first has a `unit` in the response `cols`. This causes headers in pivot tables to not be formatted correctly, and probably various other issues. ```json { "data": { "cols" : [ { "source" : "breakout", "schema_name" : "PUBLIC", "extra_info" : {}, "special_type" : null, "name" : "CREATED_AT", "fk_field_id" : null, "visibility_type" : "normal", "display_name" : "Created At", "target" : null, "base_type" : "type/DateTime", "id" : 1, "description" : "The date and time an order was submitted.", "table_id" : 1, "unit" : "month" }, { "preview_display" : true, "display_name" : "CREATED_AT_2", "target" : null, "id" : null, "base_type" : "type/DateTime", "table_id" : null, "description" : null, "source" : "breakout", "extra_info" : {}, "special_type" : null, "name" : "CREATED_AT_2" }, { "extra_info" : {}, "special_type" : "type/Number", "name" : "count", "source" : "aggregation", "base_type" : "type/Integer", "id" : null, "description" : null, "table_id" : null, "target" : null, "display_name" : "count" } ] }, "json_query" : { "query" : { "source_table" : 1, "aggregation" : [ [ "count" ] ], "breakout" : [ [ "datetime-field", [ "field-id", 1 ], "as", "month" ], [ "datetime-field", [ "fk->", 3, 22 ], "as", "month" ] ] }, "type" : "query", "parameters" : [], "database" : 1, "constraints" : { "max-results-bare-rows" : 2000, "max-results" : 10000 } } } ```
process
second date breakout doesn t have unit in the result col if i breakout by date columns only the first has a unit in the response cols this causes headers in pivot tables to not be formatted correctly and probably various other issues json data cols source breakout schema name public extra info special type null name created at fk field id null visibility type normal display name created at target null base type type datetime id description the date and time an order was submitted table id unit month preview display true display name created at target null id null base type type datetime table id null description null source breakout extra info special type null name created at extra info special type type number name count source aggregation base type type integer id null description null table id null target null display name count json query query source table aggregation count breakout datetime field field id as month datetime field fk as month type query parameters database constraints max results bare rows max results
1
2,284
5,108,914,081
IssuesEvent
2017-01-05 19:10:51
jlm2017/jlm-video-subtitles
https://api.github.com/repos/jlm2017/jlm-video-subtitles
opened
[Subtitles] [FR] VOEUX DE JEAN-LUC MÉLENCHON - UNE FRANCE AU SERVICE DE LA PAIX
Language: French Process: Someone is working on this issue Process: [1] Writing in progress
# Video title VOEUX DE JEAN-LUC MÉLENCHON - UNE FRANCE AU SERVICE DE LA PAIX # URL https://www.youtube.com/watch?v=u8Lta7F4SgY # Youtube subtitles language Français # Duration 33:56 # Subtitles URL https://www.youtube.com/timedtext_editor?ref=player&tab=captions&lang=fr&v=u8Lta7F4SgY&action_mde_edit_form=1&ui=hd&bl=vmp
2.0
[Subtitles] [FR] VOEUX DE JEAN-LUC MÉLENCHON - UNE FRANCE AU SERVICE DE LA PAIX - # Video title VOEUX DE JEAN-LUC MÉLENCHON - UNE FRANCE AU SERVICE DE LA PAIX # URL https://www.youtube.com/watch?v=u8Lta7F4SgY # Youtube subtitles language Français # Duration 33:56 # Subtitles URL https://www.youtube.com/timedtext_editor?ref=player&tab=captions&lang=fr&v=u8Lta7F4SgY&action_mde_edit_form=1&ui=hd&bl=vmp
process
voeux de jean luc mélenchon une france au service de la paix video title voeux de jean luc mélenchon une france au service de la paix url youtube subtitles language français duration subtitles url
1
19,888
26,334,712,506
IssuesEvent
2023-01-10 13:32:08
bazelbuild/bazel
https://api.github.com/repos/bazelbuild/bazel
closed
How can I make some build args as required and fail the bazel build when the arg is not provided?
P3 type: support / not a bug (process) team-OSS
How can I make some build args as required and fail the bazel build when the arg is not provided? I have a bazel build <target> --config=X, I need a way to fail the bazel build if X is not provided.
1.0
How can I make some build args as required and fail the bazel build when the arg is not provided? - How can I make some build args as required and fail the bazel build when the arg is not provided? I have a bazel build <target> --config=X, I need a way to fail the bazel build if X is not provided.
process
how can i make some build args as required and fail the bazel build when the arg is not provided how can i make some build args as required and fail the bazel build when the arg is not provided i have a bazel build config x i need a way to fail the bazel build if x is not provided
1
199,337
22,693,305,426
IssuesEvent
2022-07-05 01:10:38
TechnoConserve/personal_website
https://api.github.com/repos/TechnoConserve/personal_website
opened
jquery-3.0.0.min.js: 3 vulnerabilities (highest severity is: 6.1)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-3.0.0.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0/jquery.min.js</a></p> <p>Path to dependency file: /photo_blog/templates/base.html</p> <p>Path to vulnerable library: /photo_blog/templates/base.html</p> <p> </details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2020-11023](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-3.0.0.min.js | Direct | jquery - 3.5.0;jquery-rails - 4.4.0 | &#10060; | | [CVE-2020-11022](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-3.0.0.min.js | Direct | jQuery - 3.5.0 | &#10060; | | [CVE-2019-11358](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11358) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-3.0.0.min.js | Direct | 3.4.0 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-11023</summary> ### Vulnerable Library - <b>jquery-3.0.0.min.js</b></p> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0/jquery.min.js</a></p> <p>Path to dependency file: /photo_blog/templates/base.html</p> <p>Path to vulnerable library: /photo_blog/templates/base.html</p> <p> Dependency Hierarchy: - :x: **jquery-3.0.0.min.js** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023>CVE-2020-11023</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440">https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jquery - 3.5.0;jquery-rails - 4.4.0</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-11022</summary> ### Vulnerable Library - <b>jquery-3.0.0.min.js</b></p> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0/jquery.min.js</a></p> <p>Path to dependency file: /photo_blog/templates/base.html</p> <p>Path to vulnerable library: /photo_blog/templates/base.html</p> <p> Dependency Hierarchy: - :x: **jquery-3.0.0.min.js** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jQuery - 3.5.0</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2019-11358</summary> ### Vulnerable Library - <b>jquery-3.0.0.min.js</b></p> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0/jquery.min.js</a></p> <p>Path to dependency file: /photo_blog/templates/base.html</p> <p>Path to vulnerable library: /photo_blog/templates/base.html</p> <p> Dependency Hierarchy: - :x: **jquery-3.0.0.min.js** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype. <p>Publish Date: 2019-04-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11358>CVE-2019-11358</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358</a></p> <p>Release Date: 2019-04-20</p> <p>Fix Resolution: 3.4.0</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
True
jquery-3.0.0.min.js: 3 vulnerabilities (highest severity is: 6.1) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-3.0.0.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0/jquery.min.js</a></p> <p>Path to dependency file: /photo_blog/templates/base.html</p> <p>Path to vulnerable library: /photo_blog/templates/base.html</p> <p> </details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | --- | --- | | [CVE-2020-11023](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-3.0.0.min.js | Direct | jquery - 3.5.0;jquery-rails - 4.4.0 | &#10060; | | [CVE-2020-11022](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-3.0.0.min.js | Direct | jQuery - 3.5.0 | &#10060; | | [CVE-2019-11358](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11358) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.1 | jquery-3.0.0.min.js | Direct | 3.4.0 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-11023</summary> ### Vulnerable Library - <b>jquery-3.0.0.min.js</b></p> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0/jquery.min.js</a></p> <p>Path to dependency file: /photo_blog/templates/base.html</p> <p>Path to vulnerable library: /photo_blog/templates/base.html</p> <p> Dependency Hierarchy: - :x: **jquery-3.0.0.min.js** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023>CVE-2020-11023</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440">https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jquery - 3.5.0;jquery-rails - 4.4.0</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2020-11022</summary> ### Vulnerable Library - <b>jquery-3.0.0.min.js</b></p> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0/jquery.min.js</a></p> <p>Path to dependency file: /photo_blog/templates/base.html</p> <p>Path to vulnerable library: /photo_blog/templates/base.html</p> <p> Dependency Hierarchy: - :x: **jquery-3.0.0.min.js** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jQuery - 3.5.0</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2019-11358</summary> ### Vulnerable Library - <b>jquery-3.0.0.min.js</b></p> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/3.0.0/jquery.min.js</a></p> <p>Path to dependency file: /photo_blog/templates/base.html</p> <p>Path to vulnerable library: /photo_blog/templates/base.html</p> <p> Dependency Hierarchy: - :x: **jquery-3.0.0.min.js** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype. <p>Publish Date: 2019-04-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11358>CVE-2019-11358</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.1</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358</a></p> <p>Release Date: 2019-04-20</p> <p>Fix Resolution: 3.4.0</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
non_process
jquery min js vulnerabilities highest severity is vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file photo blog templates base html path to vulnerable library photo blog templates base html vulnerabilities cve severity cvss dependency type fixed in remediation available medium jquery min js direct jquery jquery rails medium jquery min js direct jquery medium jquery min js direct details cve vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file photo blog templates base html path to vulnerable library photo blog templates base html dependency hierarchy x jquery min js vulnerable library found in base branch main vulnerability details in jquery versions greater than or equal to and before passing html containing elements from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery jquery rails step up your open source security game with mend cve vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file photo blog templates base html path to vulnerable library photo blog templates base html dependency hierarchy x jquery min js vulnerable library found in base branch main vulnerability details in jquery versions greater than or equal to and before passing html from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with mend cve vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file photo blog templates base html path to vulnerable library photo blog templates base html dependency hierarchy x jquery min js vulnerable library found in base branch main vulnerability details jquery before as used in drupal backdrop cms and other products mishandles jquery extend true because of object prototype pollution if an unsanitized source object contained an enumerable proto property it could extend the native object prototype publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
55,683
11,458,267,115
IssuesEvent
2020-02-07 02:43:28
joomla/joomla-cms
https://api.github.com/repos/joomla/joomla-cms
closed
Error in installing Hindi language extension.
No Code Attached Yet
### Steps to reproduce the issue Go to system settings and try to install hindi language extenstion. ### Expected result ### Actual result ![Screenshot (54)](https://user-images.githubusercontent.com/55142272/73995888-1a0e2e80-4980-11ea-9ea0-d9c0472f3226.png) ### System information (as much as possible) version: 4.0.0-beta1-dev ### Additional comments
1.0
Error in installing Hindi language extension. - ### Steps to reproduce the issue Go to system settings and try to install hindi language extenstion. ### Expected result ### Actual result ![Screenshot (54)](https://user-images.githubusercontent.com/55142272/73995888-1a0e2e80-4980-11ea-9ea0-d9c0472f3226.png) ### System information (as much as possible) version: 4.0.0-beta1-dev ### Additional comments
non_process
error in installing hindi language extension steps to reproduce the issue go to system settings and try to install hindi language extenstion expected result actual result system information as much as possible version dev additional comments
0
39,735
20,177,619,101
IssuesEvent
2022-02-10 15:41:44
JuliaRobotics/IncrementalInference.jl
https://api.github.com/repos/JuliaRobotics/IncrementalInference.jl
closed
Must add cache to CalcFactor
performance types factors
Must add cache to CalcFactor _Originally posted by @dehann in https://github.com/JuliaRobotics/IncrementalInference.jl/issues/1462#issuecomment-1008447079_
True
Must add cache to CalcFactor - Must add cache to CalcFactor _Originally posted by @dehann in https://github.com/JuliaRobotics/IncrementalInference.jl/issues/1462#issuecomment-1008447079_
non_process
must add cache to calcfactor must add cache to calcfactor originally posted by dehann in
0
19,649
26,008,904,910
IssuesEvent
2022-12-20 22:30:44
CSE201-project/PaperFriend-desktop-app
https://api.github.com/repos/CSE201-project/PaperFriend-desktop-app
closed
Load from file/ save to file activities and friends
file processing
As you can see at the top of mainwindow.cpp, there are a few vectors that are going to be shared by many components of the app. Here are the two you have to focus on: std::vector<Activity>vector_activities; std::vector<Friend>vector_friends; They represent all the activities resp. friends that the user can select when writing an entry. - When the app is launched, you should load form a file activities.json and friends.json the activities and friends that are in these two vectors. - When the app is closed, you have to save back the current content of these two vectors in friends.json and activities.json
1.0
Load from file/ save to file activities and friends - As you can see at the top of mainwindow.cpp, there are a few vectors that are going to be shared by many components of the app. Here are the two you have to focus on: std::vector<Activity>vector_activities; std::vector<Friend>vector_friends; They represent all the activities resp. friends that the user can select when writing an entry. - When the app is launched, you should load form a file activities.json and friends.json the activities and friends that are in these two vectors. - When the app is closed, you have to save back the current content of these two vectors in friends.json and activities.json
process
load from file save to file activities and friends as you can see at the top of mainwindow cpp there are a few vectors that are going to be shared by many components of the app here are the two you have to focus on std vector vector activities std vector vector friends they represent all the activities resp friends that the user can select when writing an entry when the app is launched you should load form a file activities json and friends json the activities and friends that are in these two vectors when the app is closed you have to save back the current content of these two vectors in friends json and activities json
1
20,948
27,807,732,079
IssuesEvent
2023-03-17 21:54:56
cse442-at-ub/project_s23-iweatherify
https://api.github.com/repos/cse442-at-ub/project_s23-iweatherify
closed
Add button navigation functionality for the Menu
Processing Task Sprint 2
**Tests** What the Menu looks like: ![Screenshot 2023-03-15 at 11.17.23 PM.png](https://images.zenhubusercontent.com/63e16e3b1c90f75dc8056512/27067f22-b609-48ff-8bd5-a325e5bb5e33) *Test 1* 1) Go to https://github.com/cse442-at-ub/project_s23-iweatherify/tree/55_menu_button_nav 2) Click on <> Code 3) Download as a Zip 4) Follow the instructions in README.md on how to start and run on localhost 5) Observe that as a non-logged-in user on the homepage, the menu can be opened/closed by pressing the hamburger menu ![Screenshot 2023-03-15 at 11.01.21 PM.png] *Test 2* 1) Follow Test 1 Up until (4) 2) Observe that as a logged-in user on the homepage, the menu can be opened/closed by pressing the hamburger menu ![Screenshot 2023-03-15 at 11.01.21 PM.png] *Test 2* 1) Follow Test 1 Up until (4) 2) From the homepage, click on the "register" button, it should take you to the register page 3) Confirm that the register button on the current registration page is not functional and is grayed out. 4) Return to the Homepage by clicking on that button and confirm it takes you to the non-logged-in homepage 5) From the homepage, click on the "login" button, it should take you to the login page 6) Confirm that the login button on the current login page is not functional and is grayed out. 7) Return to the Homepage by clicking on that button and confirm it takes you to the non-logged-in homepage *Test 3* 1) Follow Test 1 Up until (4) 2) From the Register Page, click on the "login" button, it should take you to the login page 3) From the Login Page, click on the "Register" button it should take you to the register page *Test 4* 1) Follow Test 1 Up until (4) 2) Go to the login page using the button navigation 3) Enter the username "zal" and password "kal" and press Login 4) Confirm that The Homepage button is grayed out 5) Click on the "Unit Settings" button, it should take you to the Unit Settings Page. 6) Open the Menu Once More and Click on Homepage confirming it's working 7) From the Homepage click on "Temperature Settings" confirming that it takes you to the temperature settings page 8) Click on the Homepage to return *Test 5* 1) Follow Test 4 up until (4) 2) Open the menu and click on the Unit Settings button to take you to the Unit Settings Page 3) From this page click on "Temperature Settings", it should take you to the appropriate page 4) From the "Temperature Settings" Page open the menu and click on "Unit Settings", it should take you to the unit settings page *Test 6* 1) Follow Test 4 up until (4) 2) Click on the logout button, confirming it logs you out 3) Log back into your account with "zal" as the username "kal" as the password and login 4) Navigate to the Temperature Settings Page and click on log out from that page 5) Log back into your account with "zal" as the username "kal" as the password and login 6) Navigate to the Unit Settings Page and click on log out from that page 7) Confirm that Log Out Functions Properly ![Screenshot 2023-03-15 at 11.17.02 PM.png](https://images.zenhubusercontent.com/63e16e3b1c90f75dc8056512/8fd63e9d-eeb0-4f8c-ac60-97c6df9a0686)
1.0
Add button navigation functionality for the Menu - **Tests** What the Menu looks like: ![Screenshot 2023-03-15 at 11.17.23 PM.png](https://images.zenhubusercontent.com/63e16e3b1c90f75dc8056512/27067f22-b609-48ff-8bd5-a325e5bb5e33) *Test 1* 1) Go to https://github.com/cse442-at-ub/project_s23-iweatherify/tree/55_menu_button_nav 2) Click on <> Code 3) Download as a Zip 4) Follow the instructions in README.md on how to start and run on localhost 5) Observe that as a non-logged-in user on the homepage, the menu can be opened/closed by pressing the hamburger menu ![Screenshot 2023-03-15 at 11.01.21 PM.png] *Test 2* 1) Follow Test 1 Up until (4) 2) Observe that as a logged-in user on the homepage, the menu can be opened/closed by pressing the hamburger menu ![Screenshot 2023-03-15 at 11.01.21 PM.png] *Test 2* 1) Follow Test 1 Up until (4) 2) From the homepage, click on the "register" button, it should take you to the register page 3) Confirm that the register button on the current registration page is not functional and is grayed out. 4) Return to the Homepage by clicking on that button and confirm it takes you to the non-logged-in homepage 5) From the homepage, click on the "login" button, it should take you to the login page 6) Confirm that the login button on the current login page is not functional and is grayed out. 7) Return to the Homepage by clicking on that button and confirm it takes you to the non-logged-in homepage *Test 3* 1) Follow Test 1 Up until (4) 2) From the Register Page, click on the "login" button, it should take you to the login page 3) From the Login Page, click on the "Register" button it should take you to the register page *Test 4* 1) Follow Test 1 Up until (4) 2) Go to the login page using the button navigation 3) Enter the username "zal" and password "kal" and press Login 4) Confirm that The Homepage button is grayed out 5) Click on the "Unit Settings" button, it should take you to the Unit Settings Page. 6) Open the Menu Once More and Click on Homepage confirming it's working 7) From the Homepage click on "Temperature Settings" confirming that it takes you to the temperature settings page 8) Click on the Homepage to return *Test 5* 1) Follow Test 4 up until (4) 2) Open the menu and click on the Unit Settings button to take you to the Unit Settings Page 3) From this page click on "Temperature Settings", it should take you to the appropriate page 4) From the "Temperature Settings" Page open the menu and click on "Unit Settings", it should take you to the unit settings page *Test 6* 1) Follow Test 4 up until (4) 2) Click on the logout button, confirming it logs you out 3) Log back into your account with "zal" as the username "kal" as the password and login 4) Navigate to the Temperature Settings Page and click on log out from that page 5) Log back into your account with "zal" as the username "kal" as the password and login 6) Navigate to the Unit Settings Page and click on log out from that page 7) Confirm that Log Out Functions Properly ![Screenshot 2023-03-15 at 11.17.02 PM.png](https://images.zenhubusercontent.com/63e16e3b1c90f75dc8056512/8fd63e9d-eeb0-4f8c-ac60-97c6df9a0686)
process
add button navigation functionality for the menu tests what the menu looks like test go to click on code download as a zip follow the instructions in readme md on how to start and run on localhost observe that as a non logged in user on the homepage the menu can be opened closed by pressing the hamburger menu test follow test up until observe that as a logged in user on the homepage the menu can be opened closed by pressing the hamburger menu test follow test up until from the homepage click on the register button it should take you to the register page confirm that the register button on the current registration page is not functional and is grayed out return to the homepage by clicking on that button and confirm it takes you to the non logged in homepage from the homepage click on the login button it should take you to the login page confirm that the login button on the current login page is not functional and is grayed out return to the homepage by clicking on that button and confirm it takes you to the non logged in homepage test follow test up until from the register page click on the login button it should take you to the login page from the login page click on the register button it should take you to the register page test follow test up until go to the login page using the button navigation enter the username zal and password kal and press login confirm that the homepage button is grayed out click on the unit settings button it should take you to the unit settings page open the menu once more and click on homepage confirming it s working from the homepage click on temperature settings confirming that it takes you to the temperature settings page click on the homepage to return test follow test up until open the menu and click on the unit settings button to take you to the unit settings page from this page click on temperature settings it should take you to the appropriate page from the temperature settings page open the menu and click on unit settings it should take you to the unit settings page test follow test up until click on the logout button confirming it logs you out log back into your account with zal as the username kal as the password and login navigate to the temperature settings page and click on log out from that page log back into your account with zal as the username kal as the password and login navigate to the unit settings page and click on log out from that page confirm that log out functions properly
1
1,113
3,590,326,369
IssuesEvent
2016-02-01 04:43:19
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
ProcessTests.TestWorkingSet64 failed on OS X in CI
System.Diagnostics.Process
http://dotnet-ci.cloudapp.net/job/dotnet_corefx_mac_debug_tst/31/console ``` 21:51:43 System.Diagnostics.ProcessTests.ProcessTests.TestWorkingSet64 [FAIL] 21:51:43 Assert+WrapperXunitException : File path: d:\j\workspace\dotnet_corefx_windows_debug\src\System.Diagnostics.Process\tests\ProcessTests.cs. Line: 339 21:51:43 ---- Assert.True() Failure 21:51:43 Expected: True 21:51:43 Actual: False 21:51:43 Stack Trace: 21:51:43 at Assert.WrapException(Exception e, String callerFilePath, Int32 callerLineNumber) 21:51:43 at Assert.True(Boolean condition, String userMessage, String path, Int32 line) 21:51:43 at System.Diagnostics.ProcessTests.ProcessTests.TestWorkingSet64() 21:51:43 ----- Inner Stack Trace ----- 21:51:43 at Assert.True(Boolean condition, String userMessage, String path, Int32 line) ```
1.0
ProcessTests.TestWorkingSet64 failed on OS X in CI - http://dotnet-ci.cloudapp.net/job/dotnet_corefx_mac_debug_tst/31/console ``` 21:51:43 System.Diagnostics.ProcessTests.ProcessTests.TestWorkingSet64 [FAIL] 21:51:43 Assert+WrapperXunitException : File path: d:\j\workspace\dotnet_corefx_windows_debug\src\System.Diagnostics.Process\tests\ProcessTests.cs. Line: 339 21:51:43 ---- Assert.True() Failure 21:51:43 Expected: True 21:51:43 Actual: False 21:51:43 Stack Trace: 21:51:43 at Assert.WrapException(Exception e, String callerFilePath, Int32 callerLineNumber) 21:51:43 at Assert.True(Boolean condition, String userMessage, String path, Int32 line) 21:51:43 at System.Diagnostics.ProcessTests.ProcessTests.TestWorkingSet64() 21:51:43 ----- Inner Stack Trace ----- 21:51:43 at Assert.True(Boolean condition, String userMessage, String path, Int32 line) ```
process
processtests failed on os x in ci system diagnostics processtests processtests assert wrapperxunitexception file path d j workspace dotnet corefx windows debug src system diagnostics process tests processtests cs line assert true failure expected true actual false stack trace at assert wrapexception exception e string callerfilepath callerlinenumber at assert true boolean condition string usermessage string path line at system diagnostics processtests processtests inner stack trace at assert true boolean condition string usermessage string path line
1
380,132
26,403,897,084
IssuesEvent
2023-01-13 05:33:14
tcmetzger/sphinx-favicon
https://api.github.com/repos/tcmetzger/sphinx-favicon
closed
do you want to create a sphinx single page documentation ?
documentation enhancement
I saw sphinx in the dev-requirements.txt file and I was curious if you wanted to use it ?
1.0
do you want to create a sphinx single page documentation ? - I saw sphinx in the dev-requirements.txt file and I was curious if you wanted to use it ?
non_process
do you want to create a sphinx single page documentation i saw sphinx in the dev requirements txt file and i was curious if you wanted to use it
0
367,298
10,851,789,829
IssuesEvent
2019-11-13 11:28:47
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.google.com - desktop site instead of mobile site
browser-firefox-tablet engine-gecko priority-critical
<!-- @browser: Firefox Mobile (Tablet) 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Tablet; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: mobile-reporter --> **URL**: https://www.google.com/#spf=1573546776401 **Browser / Version**: Firefox Mobile (Tablet) 68.0 **Operating System**: Android 8.1.0 **Tested Another Browser**: Yes **Problem type**: Desktop site instead of mobile site **Description**: when clicking on image in Google, site appears instead of photo **Steps to Reproduce**: [![Screenshot Description](https://webcompat.com/uploads/2019/11/9931b5f6-c80c-4174-90b2-519f2d2c0997-thumb.jpeg)](https://webcompat.com/uploads/2019/11/9931b5f6-c80c-4174-90b2-519f2d2c0997.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20191017184803</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.google.com - desktop site instead of mobile site - <!-- @browser: Firefox Mobile (Tablet) 68.0 --> <!-- @ua_header: Mozilla/5.0 (Android 8.1.0; Tablet; rv:68.0) Gecko/68.0 Firefox/68.0 --> <!-- @reported_with: mobile-reporter --> **URL**: https://www.google.com/#spf=1573546776401 **Browser / Version**: Firefox Mobile (Tablet) 68.0 **Operating System**: Android 8.1.0 **Tested Another Browser**: Yes **Problem type**: Desktop site instead of mobile site **Description**: when clicking on image in Google, site appears instead of photo **Steps to Reproduce**: [![Screenshot Description](https://webcompat.com/uploads/2019/11/9931b5f6-c80c-4174-90b2-519f2d2c0997-thumb.jpeg)](https://webcompat.com/uploads/2019/11/9931b5f6-c80c-4174-90b2-519f2d2c0997.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20191017184803</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
desktop site instead of mobile site url browser version firefox mobile tablet operating system android tested another browser yes problem type desktop site instead of mobile site description when clicking on image in google site appears instead of photo steps to reproduce browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
0
19,529
25,840,079,329
IssuesEvent
2022-12-12 23:15:30
OpenDataScotland/the_od_bods
https://api.github.com/repos/OpenDataScotland/the_od_bods
opened
Extract category keywords from dataset title and description
data processing back end
**Is your feature request related to a problem? Please describe.** Too many datasets are set as uncategorised. This is because current categorisation uses category keywords provided by the publisher. Where there are no keywords provided by the publisher, the dataset cannot be further categorised by us in the context of the ODS catalogue. There is a similar ticket #172 but it is a large ticket to tackle. This ticket is one step down, a subset just to extract keywords out of the dataset title and description to use for categorisation, categorisation still using the existing keyword matching system. **Describe the solution you'd like** - Combine dataset title and description into single string/ text body - tokenise and remove stopwords - for each remaining keyword, get frequency count in body and return a matching category - retain top 5 most common categories (based on frequency counts) and set as dataset categories **Describe alternatives you've considered** - consider standardising casing and stemming for more accurate comparison - consider that there may be no need to cap the number of categories - i.e. returning all categories may be appropriate. See this as a % of total categories - if most categories return most of the time, then it's a meaningless solution. But because we manually curate category keywords, it might actually be filtered enough. - If the resulting categorisation takes too long (compared to current merge_data.py performance) then consider categorising on top n keywords only (instead of all keywords). The catch is that the top common words may not be useful keywords, but we may be satisfied enough with speed and small % of datasets left uncategorised. - consider TF-IDF principles (although may be more appropriate at #172 stage) **Additional context** Completion of this ticket leaves #172 to be an exploratory piece using unsupervised learning, but still a step-up in performance until then.
1.0
Extract category keywords from dataset title and description - **Is your feature request related to a problem? Please describe.** Too many datasets are set as uncategorised. This is because current categorisation uses category keywords provided by the publisher. Where there are no keywords provided by the publisher, the dataset cannot be further categorised by us in the context of the ODS catalogue. There is a similar ticket #172 but it is a large ticket to tackle. This ticket is one step down, a subset just to extract keywords out of the dataset title and description to use for categorisation, categorisation still using the existing keyword matching system. **Describe the solution you'd like** - Combine dataset title and description into single string/ text body - tokenise and remove stopwords - for each remaining keyword, get frequency count in body and return a matching category - retain top 5 most common categories (based on frequency counts) and set as dataset categories **Describe alternatives you've considered** - consider standardising casing and stemming for more accurate comparison - consider that there may be no need to cap the number of categories - i.e. returning all categories may be appropriate. See this as a % of total categories - if most categories return most of the time, then it's a meaningless solution. But because we manually curate category keywords, it might actually be filtered enough. - If the resulting categorisation takes too long (compared to current merge_data.py performance) then consider categorising on top n keywords only (instead of all keywords). The catch is that the top common words may not be useful keywords, but we may be satisfied enough with speed and small % of datasets left uncategorised. - consider TF-IDF principles (although may be more appropriate at #172 stage) **Additional context** Completion of this ticket leaves #172 to be an exploratory piece using unsupervised learning, but still a step-up in performance until then.
process
extract category keywords from dataset title and description is your feature request related to a problem please describe too many datasets are set as uncategorised this is because current categorisation uses category keywords provided by the publisher where there are no keywords provided by the publisher the dataset cannot be further categorised by us in the context of the ods catalogue there is a similar ticket but it is a large ticket to tackle this ticket is one step down a subset just to extract keywords out of the dataset title and description to use for categorisation categorisation still using the existing keyword matching system describe the solution you d like combine dataset title and description into single string text body tokenise and remove stopwords for each remaining keyword get frequency count in body and return a matching category retain top most common categories based on frequency counts and set as dataset categories describe alternatives you ve considered consider standardising casing and stemming for more accurate comparison consider that there may be no need to cap the number of categories i e returning all categories may be appropriate see this as a of total categories if most categories return most of the time then it s a meaningless solution but because we manually curate category keywords it might actually be filtered enough if the resulting categorisation takes too long compared to current merge data py performance then consider categorising on top n keywords only instead of all keywords the catch is that the top common words may not be useful keywords but we may be satisfied enough with speed and small of datasets left uncategorised consider tf idf principles although may be more appropriate at stage additional context completion of this ticket leaves to be an exploratory piece using unsupervised learning but still a step up in performance until then
1
10,136
13,044,162,426
IssuesEvent
2020-07-29 03:47:32
tikv/tikv
https://api.github.com/repos/tikv/tikv
closed
UCP: Migrate scalar function `JsonPrettySig` from TiDB
challenge-program-2 component/coprocessor difficulty/easy sig/coprocessor
## Description Port the scalar function `JsonPrettySig` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @breeswish ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
2.0
UCP: Migrate scalar function `JsonPrettySig` from TiDB - ## Description Port the scalar function `JsonPrettySig` from TiDB to coprocessor. ## Score * 50 ## Mentor(s) * @breeswish ## Recommended Skills * Rust programming ## Learning Materials Already implemented expressions ported from TiDB - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/rpn_expr) - https://github.com/tikv/tikv/tree/master/components/tidb_query/src/expr)
process
ucp migrate scalar function jsonprettysig from tidb description port the scalar function jsonprettysig from tidb to coprocessor score mentor s breeswish recommended skills rust programming learning materials already implemented expressions ported from tidb
1
18,649
24,581,037,496
IssuesEvent
2022-10-13 15:37:42
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[FHIR] QuestionnaireResponse > JSON file > Response value is not getting mapped in the following scenario
Bug P0 Response datastore Process: Fixed Process: Tested dev
Steps: 1. In SB create a study with Text choice response type activity by using 'Other' answer options 2. launch the study 3. In mobile, Sign up or sign in to the mobile app 4. Enroll to the above study 5. Submit the response for above created activity by selecting Other option 6. Observe AR: QuestionnaireResponse > JSON file > Response value is not getting mapped ER: QuestionnaireResponse > JSON file > Response value should get mapped into JSON even though when the participant selects the Other option ![image](https://user-images.githubusercontent.com/71445210/185109730-22b1851e-89c0-4ba0-983a-e1676b1afe9b.png)
2.0
[FHIR] QuestionnaireResponse > JSON file > Response value is not getting mapped in the following scenario - Steps: 1. In SB create a study with Text choice response type activity by using 'Other' answer options 2. launch the study 3. In mobile, Sign up or sign in to the mobile app 4. Enroll to the above study 5. Submit the response for above created activity by selecting Other option 6. Observe AR: QuestionnaireResponse > JSON file > Response value is not getting mapped ER: QuestionnaireResponse > JSON file > Response value should get mapped into JSON even though when the participant selects the Other option ![image](https://user-images.githubusercontent.com/71445210/185109730-22b1851e-89c0-4ba0-983a-e1676b1afe9b.png)
process
questionnaireresponse json file response value is not getting mapped in the following scenario steps in sb create a study with text choice response type activity by using other answer options launch the study in mobile sign up or sign in to the mobile app enroll to the above study submit the response for above created activity by selecting other option observe ar questionnaireresponse json file response value is not getting mapped er questionnaireresponse json file response value should get mapped into json even though when the participant selects the other option
1
29,583
14,189,826,347
IssuesEvent
2020-11-14 02:40:44
timotheecour/Nim
https://api.github.com/repos/timotheecour/Nim
closed
--gc:arc 10_000 times slower
performance
from https://github.com/nim-lang/Nim/pull/15903#discussion_r523346150 nim r -d:danger -d:case1 --hints:off $timn_D/tests/nim/all/t11283.nim OK! isNumeric3() time consuming:1.999999999999832e-06 nim r --gc:arc -d:danger -d:case1 --hints:off $timn_D/tests/nim/all/t11283.nim OK! isNumeric3() time consuming:0.010939 ```nim import times func isNumeric3*(s: string, enableNaNInf, enableLooseDot = false): bool = ## Checks whether the string is numeric. ## When the string is an integer, float or exponential, it returns true, ## otherwise it returns false. ## The `enableNaNInf` value indicates whether the `NaN` and `Inf` values ## are valid. ## The `enableLooseDot` value indicates whether loose point values such as ## `.9` and `9.` are valid. ## ## **Note:** The reason that `parseFloat()` is not used to achieve this is its ## poor performance. let length = s.len if length == 3: if (s[0] in {'i', 'I'} and s[1] in {'n', 'N'} and s[2] in {'f', 'F'}) or (s[0] in {'n', 'N'} and s[1] in {'a', 'A'} and s[2] in {'n', 'N'}): if enableNaNInf: return true else: return false if length == 4: if (s[0] in {'+', '-'} and s[1] in {'i', 'I'} and s[2] in {'n', 'N'} and s[3] in {'f', 'F'}) or (s[0] in {'+', '-'} and s[1] in {'n', 'N'} and s[2] in {'a', 'A'} and s[3] in {'n', 'N'}): if enableNaNInf: return true else: return false var eLeft, eRight, dot, e, num = false let sHigh = s.len - 1 for i, si in s: case si of '+', '-': if i == sHigh: return false if e == false: if num: return false if eLeft: return false eLeft = true else: if num: return false if eRight: return false eRight = true of '.': if dot: return false if not enableLooseDot: if num == false: return false else: if i == sHigh: return false if s[i+1] in {'e', 'E'}: return false num = false dot = true of 'e', 'E': if i == sHigh: return false if num or dot: if e: return false num = false else: return false e = true of '0'..'9': num = true of '_': if num == false: return false if dot and num == false: return false if e and num == false: return false else: return false return true proc testIsNumeric3() = var a = 0 let execptedValue = 160000000 let t0 = cputime() for i in 0..9999999: if isNumeric3("123"): a += 1 if isNumeric3("123.45"): a += 1 if isNumeric3("+123.45"): a += 1 if isNumeric3("123.45e-2"): a += 1 if isNumeric3("+123.45E-2"): a += 1 if isNumeric3("-123.45e2"): a += 1 if not isNumeric3("e123.45"): a += 1 if not isNumeric3("abc"): a += 1 if not isNumeric3("123abc"): a += 1 if not isNumeric3("123.45.6"): a += 1 if not isNumeric3("123.45e++5"): a += 1 if not isNumeric3("5.2+e1"): a += 1 if not isNumeric3(".9"): a += 1 if not isNumeric3("Inf"): a += 1 if not isNumeric3("-Inf"): a += 1 if not isNumeric3("NaN"): a += 1 let t1 = cputime() - t0 if a == execptedValue: echo "OK! isNumeric3() time consuming:", t1 testIsNumeric3() ```
True
--gc:arc 10_000 times slower - from https://github.com/nim-lang/Nim/pull/15903#discussion_r523346150 nim r -d:danger -d:case1 --hints:off $timn_D/tests/nim/all/t11283.nim OK! isNumeric3() time consuming:1.999999999999832e-06 nim r --gc:arc -d:danger -d:case1 --hints:off $timn_D/tests/nim/all/t11283.nim OK! isNumeric3() time consuming:0.010939 ```nim import times func isNumeric3*(s: string, enableNaNInf, enableLooseDot = false): bool = ## Checks whether the string is numeric. ## When the string is an integer, float or exponential, it returns true, ## otherwise it returns false. ## The `enableNaNInf` value indicates whether the `NaN` and `Inf` values ## are valid. ## The `enableLooseDot` value indicates whether loose point values such as ## `.9` and `9.` are valid. ## ## **Note:** The reason that `parseFloat()` is not used to achieve this is its ## poor performance. let length = s.len if length == 3: if (s[0] in {'i', 'I'} and s[1] in {'n', 'N'} and s[2] in {'f', 'F'}) or (s[0] in {'n', 'N'} and s[1] in {'a', 'A'} and s[2] in {'n', 'N'}): if enableNaNInf: return true else: return false if length == 4: if (s[0] in {'+', '-'} and s[1] in {'i', 'I'} and s[2] in {'n', 'N'} and s[3] in {'f', 'F'}) or (s[0] in {'+', '-'} and s[1] in {'n', 'N'} and s[2] in {'a', 'A'} and s[3] in {'n', 'N'}): if enableNaNInf: return true else: return false var eLeft, eRight, dot, e, num = false let sHigh = s.len - 1 for i, si in s: case si of '+', '-': if i == sHigh: return false if e == false: if num: return false if eLeft: return false eLeft = true else: if num: return false if eRight: return false eRight = true of '.': if dot: return false if not enableLooseDot: if num == false: return false else: if i == sHigh: return false if s[i+1] in {'e', 'E'}: return false num = false dot = true of 'e', 'E': if i == sHigh: return false if num or dot: if e: return false num = false else: return false e = true of '0'..'9': num = true of '_': if num == false: return false if dot and num == false: return false if e and num == false: return false else: return false return true proc testIsNumeric3() = var a = 0 let execptedValue = 160000000 let t0 = cputime() for i in 0..9999999: if isNumeric3("123"): a += 1 if isNumeric3("123.45"): a += 1 if isNumeric3("+123.45"): a += 1 if isNumeric3("123.45e-2"): a += 1 if isNumeric3("+123.45E-2"): a += 1 if isNumeric3("-123.45e2"): a += 1 if not isNumeric3("e123.45"): a += 1 if not isNumeric3("abc"): a += 1 if not isNumeric3("123abc"): a += 1 if not isNumeric3("123.45.6"): a += 1 if not isNumeric3("123.45e++5"): a += 1 if not isNumeric3("5.2+e1"): a += 1 if not isNumeric3(".9"): a += 1 if not isNumeric3("Inf"): a += 1 if not isNumeric3("-Inf"): a += 1 if not isNumeric3("NaN"): a += 1 let t1 = cputime() - t0 if a == execptedValue: echo "OK! isNumeric3() time consuming:", t1 testIsNumeric3() ```
non_process
gc arc times slower from nim r d danger d hints off timn d tests nim all nim ok time consuming nim r gc arc d danger d hints off timn d tests nim all nim ok time consuming nim import times func s string enablenaninf enableloosedot false bool checks whether the string is numeric when the string is an integer float or exponential it returns true otherwise it returns false the enablenaninf value indicates whether the nan and inf values are valid the enableloosedot value indicates whether loose point values such as and are valid note the reason that parsefloat is not used to achieve this is its poor performance let length s len if length if s in i i and s in n n and s in f f or s in n n and s in a a and s in n n if enablenaninf return true else return false if length if s in and s in i i and s in n n and s in f f or s in and s in n n and s in a a and s in n n if enablenaninf return true else return false var eleft eright dot e num false let shigh s len for i si in s case si of if i shigh return false if e false if num return false if eleft return false eleft true else if num return false if eright return false eright true of if dot return false if not enableloosedot if num false return false else if i shigh return false if s in e e return false num false dot true of e e if i shigh return false if num or dot if e return false num false else return false e true of num true of if num false return false if dot and num false return false if e and num false return false else return false return true proc var a let execptedvalue let cputime for i in if a if a if a if a if a if a if not a if not abc a if not a if not a if not a if not a if not a if not inf a if not inf a if not nan a let cputime if a execptedvalue echo ok time consuming
0
4,334
7,242,199,103
IssuesEvent
2018-02-14 06:16:38
muflihun/residue
https://api.github.com/repos/muflihun/residue
closed
Deadlock detected while dispatching log
area: log-processing edge-case type: bug
``` 19:03:02,606 [LogHandler] [vTRACE] [session.cc:86] Adding bytes 19:03:02,608 [LogDispatcher] [vDEBUG] [log-request-handler.cc:88] Pausing schedule for client integrity 19:03:02,609 [LogHandler] [vTRACE] [session.cc:75] Received: 522 bytes 19:03:02,609 [LogHandler] [vTRACE] [session.cc:102] Read bytes: [..snip] [size: 522] 19:03:02,609 [LogHandler] [vDEBUG] [session.cc:171] Sending {r:0} 19:03:02,609 [LogHandler] [vTRACE] [session.cc:86] Adding bytes 19:03:02,609 [LogDispatcher] [vCRAZY] [log-request-handler.cc:100] -----============= [ BEGIN ] =============----- 19:03:02,610 [LogDispatcher] [vDEBUG] [request-handler.h:144] Raw request: [..snip] 19:03:02,610 [LogDispatcher] [vDEBUG] [request-handler.cc:58] Client: [..snip] 19:03:02,610 [LogDispatcher] [vCRAZY] [request-handler.cc:59] IV: [..snip] 19:03:02,610 [LogDispatcher] [vCRAZY] [request-handler.cc:66] Data (base64): [..snip] 19:03:02,610 [LogDispatcher] [vCRAZY] [request-handler.cc:69] Ripe command: [..snip] 19:03:02,610 [LogDispatcher] [vCRAZY] [request-handler.cc:76] Plain request: [..snip] 19:03:02,610 [LogDispatcher] [vTRACE] [request-handler.h:148] Decryption finished (b64): [..snip] 19:03:02,610 [LogDispatcher] [vTRACE] [request-handler.h:159] Decompressing: [..snip] 19:03:02,611 [LogDispatcher] [vTRACE] [request-handler.h:170] Decompression finished (raw): [{... (removed)}] 19:03:02,612 [LogDispatcher] [vDEBUG] [log-request-handler.cc:115] Is bulk? true 19:03:02,612 [LogDispatcher] [vDEBUG] [log-request-handler.cc:124] Request client: 0x10065cfa8 19:03:02,612 [LogHandler] [vTRACE] [session.cc:75] Received: 522 bytes 19:03:02,612 [LogDispatcher] [vDEBUG] [log-request-handler.cc:206] Force check: true, clientRef: 0x70000f9729a0, *clientRef: muflihun00102030, bypassChecks: false 19:03:02,612 [LogHandler] [vTRACE] [session.cc:102] Read bytes: 29277885c8e3830c20fb91fab830194b:muflihun00102030:zaAqRKANlz9SElDUXg6ePbeykddpAyJjxC3TEL/yutU5HIryqLpN1xmsEncDHizpniIeaCIY7kfUjuzTaXwYJC0dxHxlssq5Lm7PzgPCyyxa7POh0quW1zooTSlBMvUiuCWamAdfzz3WPPjjbMSyvqtgMmbFIhs4qMMMt3kKZbli8kk9Hmv5XV8Ltz+/tLX4ppiOn6qSFh/pIQ0LNaBVK0wHXLfGO+AN58vc1D50JT8vA00XUjkzNNmeasekIntZDoRut/WhgqVijoUsT27sY9WzrsdaLuIJIAUhtewqvAGgyeplI+lucgHWtWKVR+HQQ887giQwW6+80FrU0yIZ/yEhDnFbCul3Zc9L3vuuBHvrAgjYvDPkA4MZPwBkFWSh6anUbJ5zsNgEnMP2tdDI+ZLnSVfvQ+ZNEPeVXYE15la/XVPoe0/b/2MqY3HPUrYEtTJONqT/XrZKl1HXbqxp4w== [size: 522] 19:03:02,613 [LogDispatcher] [vDEBUG] [client.cc:65] Checking token 37953058 (client [muflihun00102030]) ```
1.0
Deadlock detected while dispatching log - ``` 19:03:02,606 [LogHandler] [vTRACE] [session.cc:86] Adding bytes 19:03:02,608 [LogDispatcher] [vDEBUG] [log-request-handler.cc:88] Pausing schedule for client integrity 19:03:02,609 [LogHandler] [vTRACE] [session.cc:75] Received: 522 bytes 19:03:02,609 [LogHandler] [vTRACE] [session.cc:102] Read bytes: [..snip] [size: 522] 19:03:02,609 [LogHandler] [vDEBUG] [session.cc:171] Sending {r:0} 19:03:02,609 [LogHandler] [vTRACE] [session.cc:86] Adding bytes 19:03:02,609 [LogDispatcher] [vCRAZY] [log-request-handler.cc:100] -----============= [ BEGIN ] =============----- 19:03:02,610 [LogDispatcher] [vDEBUG] [request-handler.h:144] Raw request: [..snip] 19:03:02,610 [LogDispatcher] [vDEBUG] [request-handler.cc:58] Client: [..snip] 19:03:02,610 [LogDispatcher] [vCRAZY] [request-handler.cc:59] IV: [..snip] 19:03:02,610 [LogDispatcher] [vCRAZY] [request-handler.cc:66] Data (base64): [..snip] 19:03:02,610 [LogDispatcher] [vCRAZY] [request-handler.cc:69] Ripe command: [..snip] 19:03:02,610 [LogDispatcher] [vCRAZY] [request-handler.cc:76] Plain request: [..snip] 19:03:02,610 [LogDispatcher] [vTRACE] [request-handler.h:148] Decryption finished (b64): [..snip] 19:03:02,610 [LogDispatcher] [vTRACE] [request-handler.h:159] Decompressing: [..snip] 19:03:02,611 [LogDispatcher] [vTRACE] [request-handler.h:170] Decompression finished (raw): [{... (removed)}] 19:03:02,612 [LogDispatcher] [vDEBUG] [log-request-handler.cc:115] Is bulk? true 19:03:02,612 [LogDispatcher] [vDEBUG] [log-request-handler.cc:124] Request client: 0x10065cfa8 19:03:02,612 [LogHandler] [vTRACE] [session.cc:75] Received: 522 bytes 19:03:02,612 [LogDispatcher] [vDEBUG] [log-request-handler.cc:206] Force check: true, clientRef: 0x70000f9729a0, *clientRef: muflihun00102030, bypassChecks: false 19:03:02,612 [LogHandler] [vTRACE] [session.cc:102] Read bytes: 29277885c8e3830c20fb91fab830194b:muflihun00102030:zaAqRKANlz9SElDUXg6ePbeykddpAyJjxC3TEL/yutU5HIryqLpN1xmsEncDHizpniIeaCIY7kfUjuzTaXwYJC0dxHxlssq5Lm7PzgPCyyxa7POh0quW1zooTSlBMvUiuCWamAdfzz3WPPjjbMSyvqtgMmbFIhs4qMMMt3kKZbli8kk9Hmv5XV8Ltz+/tLX4ppiOn6qSFh/pIQ0LNaBVK0wHXLfGO+AN58vc1D50JT8vA00XUjkzNNmeasekIntZDoRut/WhgqVijoUsT27sY9WzrsdaLuIJIAUhtewqvAGgyeplI+lucgHWtWKVR+HQQ887giQwW6+80FrU0yIZ/yEhDnFbCul3Zc9L3vuuBHvrAgjYvDPkA4MZPwBkFWSh6anUbJ5zsNgEnMP2tdDI+ZLnSVfvQ+ZNEPeVXYE15la/XVPoe0/b/2MqY3HPUrYEtTJONqT/XrZKl1HXbqxp4w== [size: 522] 19:03:02,613 [LogDispatcher] [vDEBUG] [client.cc:65] Checking token 37953058 (client [muflihun00102030]) ```
process
deadlock detected while dispatching log adding bytes pausing schedule for client integrity received bytes read bytes sending r adding bytes raw request client iv data ripe command plain request decryption finished decompressing decompression finished raw is bulk true request client received bytes force check true clientref clientref bypasschecks false read bytes lucghwtwkvr zlnsvfvq b checking token client
1
4,580
2,734,424,272
IssuesEvent
2015-04-17 19:49:15
uonline/universe
https://api.github.com/repos/uonline/universe
opened
2.0
design discussion
Сижу вот, рисую дизайны. Придумал такой: ![screenshot from 2015-04-17 22-44-42](https://cloud.githubusercontent.com/assets/1831620/7209685/a8d32f58-e553-11e4-8354-cdf5c0561759.png) Очень нравится, в два раза удобнее нынешнего. Только непонятно, кто нам столько картинок отрисует.
1.0
2.0 - Сижу вот, рисую дизайны. Придумал такой: ![screenshot from 2015-04-17 22-44-42](https://cloud.githubusercontent.com/assets/1831620/7209685/a8d32f58-e553-11e4-8354-cdf5c0561759.png) Очень нравится, в два раза удобнее нынешнего. Только непонятно, кто нам столько картинок отрисует.
non_process
сижу вот рисую дизайны придумал такой очень нравится в два раза удобнее нынешнего только непонятно кто нам столько картинок отрисует
0
37,633
8,337,011,311
IssuesEvent
2018-09-28 09:40:14
Yoast/YoastSEO.js
https://api.github.com/repos/Yoast/YoastSEO.js
closed
Refactor SubheadingsKeyword assessment
code-review lingo morpho-syno
Refactor the subheadingsKeywordAssessment to follow the following schema: - A subheading is considered to reflect the topic if > half of content words from the keyphrase are used in it. - GOOD if 30-75% of subheadings reflect the topic, - BAD otherwise. Using synonyms: yes
1.0
Refactor SubheadingsKeyword assessment - Refactor the subheadingsKeywordAssessment to follow the following schema: - A subheading is considered to reflect the topic if > half of content words from the keyphrase are used in it. - GOOD if 30-75% of subheadings reflect the topic, - BAD otherwise. Using synonyms: yes
non_process
refactor subheadingskeyword assessment refactor the subheadingskeywordassessment to follow the following schema a subheading is considered to reflect the topic if half of content words from the keyphrase are used in it good if of subheadings reflect the topic bad otherwise using synonyms yes
0
459,927
13,201,501,732
IssuesEvent
2020-08-14 10:16:21
strapi/strapi
https://api.github.com/repos/strapi/strapi
closed
Setting a datetime in a component as an entry title causes a react error when setting the datetime
priority: low status: can't reproduce
**Describe the bug** Setting a datetime's value inside of a component using the calendar UI causes a react error if the datetime is the entry title for the component. **Steps to reproduce the behavior** (Using the admin panel UI) 1. Create a content type 2. Create a piece of content using that type 3. Create a component. 4. Add a datetime to the component 5. Go to the content created in step 2 6. Using the "edit the view" button, edit the component's layout 7. Change the entry title to the datetime added in step 4 8. Go back to edit the content created in step 2 9. In the component, attempt to set the dateTime 10. React error: https://imgur.com/a/9WYX3UY **Expected behavior** The date is set and the entry title updates accordingly. Node.js version: 12.14.1 NPM version: 6.13.4 Strapi version: Strapi v3.0.0 Database: Postgres Operating system: ubuntu 18.04.04 lts
1.0
Setting a datetime in a component as an entry title causes a react error when setting the datetime - **Describe the bug** Setting a datetime's value inside of a component using the calendar UI causes a react error if the datetime is the entry title for the component. **Steps to reproduce the behavior** (Using the admin panel UI) 1. Create a content type 2. Create a piece of content using that type 3. Create a component. 4. Add a datetime to the component 5. Go to the content created in step 2 6. Using the "edit the view" button, edit the component's layout 7. Change the entry title to the datetime added in step 4 8. Go back to edit the content created in step 2 9. In the component, attempt to set the dateTime 10. React error: https://imgur.com/a/9WYX3UY **Expected behavior** The date is set and the entry title updates accordingly. Node.js version: 12.14.1 NPM version: 6.13.4 Strapi version: Strapi v3.0.0 Database: Postgres Operating system: ubuntu 18.04.04 lts
non_process
setting a datetime in a component as an entry title causes a react error when setting the datetime describe the bug setting a datetime s value inside of a component using the calendar ui causes a react error if the datetime is the entry title for the component steps to reproduce the behavior using the admin panel ui create a content type create a piece of content using that type create a component add a datetime to the component go to the content created in step using the edit the view button edit the component s layout change the entry title to the datetime added in step go back to edit the content created in step in the component attempt to set the datetime react error expected behavior the date is set and the entry title updates accordingly node js version npm version strapi version strapi database postgres operating system ubuntu lts
0
10,948
13,756,464,341
IssuesEvent
2020-10-06 19:58:51
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Freeze and crash while trying to rename condition in conditional branch
Bug Modeller Processing
I have a Conditional branch in Model and other algorithm depends on it. If I try to rename condition branch -> Qgis freezes and crashes. ![qgis conditional branch bug](https://user-images.githubusercontent.com/874261/94351300-ff4c0f80-005f-11eb-8dd0-715d6660aa00.gif) This can be avoided though if I first turn off dependency, then rename my condition branch and turn on dependency again ![qgis conditional branch bug 2](https://user-images.githubusercontent.com/874261/94351330-5b169880-0060-11eb-9e81-31b61cae8130.gif) Sample project with model attached [Bug with conditional branch.zip](https://github.com/qgis/QGIS/files/5287404/Bug.with.conditional.branch.zip) P.s. Windows 10, Qgis 3.14 and Master
1.0
Freeze and crash while trying to rename condition in conditional branch - I have a Conditional branch in Model and other algorithm depends on it. If I try to rename condition branch -> Qgis freezes and crashes. ![qgis conditional branch bug](https://user-images.githubusercontent.com/874261/94351300-ff4c0f80-005f-11eb-8dd0-715d6660aa00.gif) This can be avoided though if I first turn off dependency, then rename my condition branch and turn on dependency again ![qgis conditional branch bug 2](https://user-images.githubusercontent.com/874261/94351330-5b169880-0060-11eb-9e81-31b61cae8130.gif) Sample project with model attached [Bug with conditional branch.zip](https://github.com/qgis/QGIS/files/5287404/Bug.with.conditional.branch.zip) P.s. Windows 10, Qgis 3.14 and Master
process
freeze and crash while trying to rename condition in conditional branch i have a conditional branch in model and other algorithm depends on it if i try to rename condition branch qgis freezes and crashes this can be avoided though if i first turn off dependency then rename my condition branch and turn on dependency again sample project with model attached p s windows qgis and master
1
9,074
12,147,383,330
IssuesEvent
2020-04-24 12:56:31
GoogleCloudPlatform/dotnet-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/dotnet-docs-samples
closed
Logging: TestListEntries is timing out.
api: logging priority: p1 type: process
Even after 10 minutes. I've skipped it in #1001 . I'll look at it.
1.0
Logging: TestListEntries is timing out. - Even after 10 minutes. I've skipped it in #1001 . I'll look at it.
process
logging testlistentries is timing out even after minutes i ve skipped it in i ll look at it
1
19,645
26,006,104,451
IssuesEvent
2022-12-20 19:32:09
openxla/stablehlo
https://api.github.com/repos/openxla/stablehlo
reopened
Nonconforming syntax in multiple markdown files
Process
### What happened? Although some MD parsers are pretty forgiving (including GitHub's), the current markdown files in stablehlo include a variety of syntax errors that do not conform to the basic [MD syntax rules](https://daringfireball.net/projects/markdown/syntax). Examples include improper indentation for lists, unescaped asterisks signs (meant for bold or italic text), and missing blank lines between headings and paragraphs. We should follow traditional MD syntax rules to ensure well-structured files that are easy to read, use an agreed-upon structure that can be tested, and that are compatible with more than one MD parser. The simplest solution is to [markdownlint](https://github.com/markdownlint/markdownlint) and fix various syntax and style issues. We can submit fixes for individual files, and make exceptions as appropriate because this tool allows you to disable certain rules in the linter. Then, once the files are cleaned up, we should enable a GitHub Action to run markdownlint on all new pull requests with .md files so we can avoid re-introduction of syntax errors. ### Steps to reproduce your issue _No response_ ### Version information _No response_
1.0
Nonconforming syntax in multiple markdown files - ### What happened? Although some MD parsers are pretty forgiving (including GitHub's), the current markdown files in stablehlo include a variety of syntax errors that do not conform to the basic [MD syntax rules](https://daringfireball.net/projects/markdown/syntax). Examples include improper indentation for lists, unescaped asterisks signs (meant for bold or italic text), and missing blank lines between headings and paragraphs. We should follow traditional MD syntax rules to ensure well-structured files that are easy to read, use an agreed-upon structure that can be tested, and that are compatible with more than one MD parser. The simplest solution is to [markdownlint](https://github.com/markdownlint/markdownlint) and fix various syntax and style issues. We can submit fixes for individual files, and make exceptions as appropriate because this tool allows you to disable certain rules in the linter. Then, once the files are cleaned up, we should enable a GitHub Action to run markdownlint on all new pull requests with .md files so we can avoid re-introduction of syntax errors. ### Steps to reproduce your issue _No response_ ### Version information _No response_
process
nonconforming syntax in multiple markdown files what happened although some md parsers are pretty forgiving including github s the current markdown files in stablehlo include a variety of syntax errors that do not conform to the basic examples include improper indentation for lists unescaped asterisks signs meant for bold or italic text and missing blank lines between headings and paragraphs we should follow traditional md syntax rules to ensure well structured files that are easy to read use an agreed upon structure that can be tested and that are compatible with more than one md parser the simplest solution is to and fix various syntax and style issues we can submit fixes for individual files and make exceptions as appropriate because this tool allows you to disable certain rules in the linter then once the files are cleaned up we should enable a github action to run markdownlint on all new pull requests with md files so we can avoid re introduction of syntax errors steps to reproduce your issue no response version information no response
1
471,208
13,562,508,671
IssuesEvent
2020-09-18 07:00:21
qlcchain/qlc-hub
https://api.github.com/repos/qlcchain/qlc-hub
closed
release v1.0
Priority: High Type: Maintenance
- [x] prepare NEO/ETH signer accounts @ariesunny - [x] update deploy scripts @gythialy - [x] tag v1.0 @zengchen221 - [ ] deploy QLC staking contract @zengchen221
1.0
release v1.0 - - [x] prepare NEO/ETH signer accounts @ariesunny - [x] update deploy scripts @gythialy - [x] tag v1.0 @zengchen221 - [ ] deploy QLC staking contract @zengchen221
non_process
release prepare neo eth signer accounts ariesunny update deploy scripts gythialy tag deploy qlc staking contract
0
16,779
21,963,281,473
IssuesEvent
2022-05-24 17:37:05
cypress-io/cypress
https://api.github.com/repos/cypress-io/cypress
closed
How to use with customize-cra
type: question npm: @cypress/webpack-preprocessor
<!-- Want a bug fixed quickly? Please provide a repository to reproduce the issue. --> How to use with `customize-cra` package.json ```json { "name": "package", "version": "1.3.0", "private": true, "scripts": { "start": "react-native start", "android": "react-native run-android", "ios": "react-native run-ios", "web": "REACT_APP_ENV=development react-app-rewired start", "build": "echo Building to ${REACT_APP_ENV} environment && react-app-rewired build", "build-dev": "REACT_APP_ENV=development CI=false npm run build", "build-prod": "REACT_APP_ENV=production CI=true npm run build", "build-android": "cd android && ./gradlew bundleRelease", "test": "jest --passWithNoTests", "cy:open": "NODE_ENV=test cypress open", "cy:run:all": "NODE_ENV=test cypress run --config video=false", "cy:run:smoke": "NODE_ENV=test cypress run --config video=false,integrationFolder=cypress/integration/smoke", "serve": "serve -s build -p 3000", "serve-test-cypress": "start-test serve 3000 cy:run:all" }, "dependencies": { "react": "16.9.0", "react-native": "0.61.5", "react-native-vector-icons": "^4.6.0", "react-native-web": "^0.11.7", "react-scripts": "^3.4.1", ... }, "devDependencies": { "@babel/core": "^7.9.0", "@babel/plugin-proposal-class-properties": "^7.8.3", "@babel/runtime": "^7.9.2", "@cypress/webpack-preprocessor": "^4.1.3", "babel-jest": "^24.9.0", "babel-plugin-react-native-web": "^0.12.2", "customize-cra": "^0.9.1", "cypress": "^4.3.0", "husky": "^4.2.3", "jest": "^24.9.0", "react-app-rewired": "^2.1.5", "react-test-renderer": "16.9.0", "serve": "^11.3.0", "start-server-and-test": "^1.10.11", ... }, "jest": { "preset": "react-native", "collectCoverageFrom": [ "src/**/*.{js,jsx,mjs}" ], "testEnvironment": "node", "testURL": "http://localhost", "moduleNameMapper": { "^react-native$": "react-native-web" }, }, "husky": { "hooks": { "pre-commit": "npm run lint" } }, "browserslist": { "production": [ ">0.2%", "not dead", "not op_mini all" ], "development": [ "last 1 chrome version", "last 1 firefox version", "last 1 safari version" ] } } ``` config-overrides.js ```javascript const { override, disableEsLint, babelInclude, addBabelPlugin, addExternalBabelPlugins, addWebpackPlugin } = require("customize-cra"); const webpack = require("webpack"); const path = require("path"); module.exports = override( disableEsLint(), addBabelPlugin("react-native-web"), addExternalBabelPlugins("@babel/plugin-proposal-class-properties"), babelInclude([ path.resolve("src"), path.resolve("node_modules/react-native-vector-icons") ]), addWebpackPlugin( new webpack.DefinePlugin({ __DEV__: process.env.REACT_APP_ENV === "development" || process.env.NODE_ENV === 'development' }) ) ); ``` - Operating System: Ubuntu 19.04 - Cypress Version: 4.3.0 - Browser Version:
1.0
How to use with customize-cra - <!-- Want a bug fixed quickly? Please provide a repository to reproduce the issue. --> How to use with `customize-cra` package.json ```json { "name": "package", "version": "1.3.0", "private": true, "scripts": { "start": "react-native start", "android": "react-native run-android", "ios": "react-native run-ios", "web": "REACT_APP_ENV=development react-app-rewired start", "build": "echo Building to ${REACT_APP_ENV} environment && react-app-rewired build", "build-dev": "REACT_APP_ENV=development CI=false npm run build", "build-prod": "REACT_APP_ENV=production CI=true npm run build", "build-android": "cd android && ./gradlew bundleRelease", "test": "jest --passWithNoTests", "cy:open": "NODE_ENV=test cypress open", "cy:run:all": "NODE_ENV=test cypress run --config video=false", "cy:run:smoke": "NODE_ENV=test cypress run --config video=false,integrationFolder=cypress/integration/smoke", "serve": "serve -s build -p 3000", "serve-test-cypress": "start-test serve 3000 cy:run:all" }, "dependencies": { "react": "16.9.0", "react-native": "0.61.5", "react-native-vector-icons": "^4.6.0", "react-native-web": "^0.11.7", "react-scripts": "^3.4.1", ... }, "devDependencies": { "@babel/core": "^7.9.0", "@babel/plugin-proposal-class-properties": "^7.8.3", "@babel/runtime": "^7.9.2", "@cypress/webpack-preprocessor": "^4.1.3", "babel-jest": "^24.9.0", "babel-plugin-react-native-web": "^0.12.2", "customize-cra": "^0.9.1", "cypress": "^4.3.0", "husky": "^4.2.3", "jest": "^24.9.0", "react-app-rewired": "^2.1.5", "react-test-renderer": "16.9.0", "serve": "^11.3.0", "start-server-and-test": "^1.10.11", ... }, "jest": { "preset": "react-native", "collectCoverageFrom": [ "src/**/*.{js,jsx,mjs}" ], "testEnvironment": "node", "testURL": "http://localhost", "moduleNameMapper": { "^react-native$": "react-native-web" }, }, "husky": { "hooks": { "pre-commit": "npm run lint" } }, "browserslist": { "production": [ ">0.2%", "not dead", "not op_mini all" ], "development": [ "last 1 chrome version", "last 1 firefox version", "last 1 safari version" ] } } ``` config-overrides.js ```javascript const { override, disableEsLint, babelInclude, addBabelPlugin, addExternalBabelPlugins, addWebpackPlugin } = require("customize-cra"); const webpack = require("webpack"); const path = require("path"); module.exports = override( disableEsLint(), addBabelPlugin("react-native-web"), addExternalBabelPlugins("@babel/plugin-proposal-class-properties"), babelInclude([ path.resolve("src"), path.resolve("node_modules/react-native-vector-icons") ]), addWebpackPlugin( new webpack.DefinePlugin({ __DEV__: process.env.REACT_APP_ENV === "development" || process.env.NODE_ENV === 'development' }) ) ); ``` - Operating System: Ubuntu 19.04 - Cypress Version: 4.3.0 - Browser Version:
process
how to use with customize cra how to use with customize cra package json json name package version private true scripts start react native start android react native run android ios react native run ios web react app env development react app rewired start build echo building to react app env environment react app rewired build build dev react app env development ci false npm run build build prod react app env production ci true npm run build build android cd android gradlew bundlerelease test jest passwithnotests cy open node env test cypress open cy run all node env test cypress run config video false cy run smoke node env test cypress run config video false integrationfolder cypress integration smoke serve serve s build p serve test cypress start test serve cy run all dependencies react react native react native vector icons react native web react scripts devdependencies babel core babel plugin proposal class properties babel runtime cypress webpack preprocessor babel jest babel plugin react native web customize cra cypress husky jest react app rewired react test renderer serve start server and test jest preset react native collectcoveragefrom src js jsx mjs testenvironment node testurl modulenamemapper react native react native web husky hooks pre commit npm run lint browserslist production not dead not op mini all development last chrome version last firefox version last safari version config overrides js javascript const override disableeslint babelinclude addbabelplugin addexternalbabelplugins addwebpackplugin require customize cra const webpack require webpack const path require path module exports override disableeslint addbabelplugin react native web addexternalbabelplugins babel plugin proposal class properties babelinclude path resolve src path resolve node modules react native vector icons addwebpackplugin new webpack defineplugin dev process env react app env development process env node env development operating system ubuntu cypress version browser version
1
86,258
8,030,274,333
IssuesEvent
2018-07-27 18:59:23
Microsoft/vscode
https://api.github.com/repos/Microsoft/vscode
opened
Test: Clickable links tsconfig.json for project references
testplan-item
Test for #54104 - [ ] Mac - [ ] Linux - [ ] Windows Complexity: 1 **Summary** With TS 3.0, tsconfig.json files may also specify project referecenes. ```js { "references": [ { "path": "./other/project/tsconfig.app.json" } ] } ``` This references field can either be a path to a json file, or a path to a folder that contains a `tsconfig.json` folder The project references paths should now be clickable links in VS code. **Test** Test that both types of links (links to tsconfigs and to folders containing tsconfigs) are clickable in VS Code and open the correct file when clicked. All paths should be resolved relative to the current tsconfig.json
1.0
Test: Clickable links tsconfig.json for project references - Test for #54104 - [ ] Mac - [ ] Linux - [ ] Windows Complexity: 1 **Summary** With TS 3.0, tsconfig.json files may also specify project referecenes. ```js { "references": [ { "path": "./other/project/tsconfig.app.json" } ] } ``` This references field can either be a path to a json file, or a path to a folder that contains a `tsconfig.json` folder The project references paths should now be clickable links in VS code. **Test** Test that both types of links (links to tsconfigs and to folders containing tsconfigs) are clickable in VS Code and open the correct file when clicked. All paths should be resolved relative to the current tsconfig.json
non_process
test clickable links tsconfig json for project references test for mac linux windows complexity summary with ts tsconfig json files may also specify project referecenes js references path other project tsconfig app json this references field can either be a path to a json file or a path to a folder that contains a tsconfig json folder the project references paths should now be clickable links in vs code test test that both types of links links to tsconfigs and to folders containing tsconfigs are clickable in vs code and open the correct file when clicked all paths should be resolved relative to the current tsconfig json
0
16,884
22,162,776,328
IssuesEvent
2022-06-04 19:04:02
maticnetwork/miden
https://api.github.com/repos/maticnetwork/miden
opened
Add operation tracking to VmStateIterator
processor
Currently, the [VmState](https://github.com/maticnetwork/miden/blob/next/processor/src/debug.rs#L7) struct which is returned from the `VmStateIterator` does not contain the operation which was executed to put the VM into this state. We should add another field to this struct so that it looks something like this: ```Rust #[derive(Clone, Debug, Eq, PartialEq)] pub struct VmState { pub op: Operation, pub clk: usize, pub fmp: Felt, pub stack: Vec<Felt>, pub memory: Vec<(u64, Word)>, } ``` The `op` field would need to be populated from the information in the [decoder](https://github.com/maticnetwork/miden/blob/next/processor/src/decoder/mod.rs#L209). The decoder doesn't explicitly track the operations yet. It is possible to infer the operations from the trace, but I think a better approach would be to have a vector of operations in the decoder struct. Then, as the VM executes operations, they would be pushed into this vector. Since tracking operations would result in some overhead, the above should happen only when we are executing programs via [execute_iter()](https://github.com/maticnetwork/miden/blob/next/processor/src/lib.rs#L80) function. Implementing this functionality well enable counting operations executed by the VM which would be useful for things like #198.
1.0
Add operation tracking to VmStateIterator - Currently, the [VmState](https://github.com/maticnetwork/miden/blob/next/processor/src/debug.rs#L7) struct which is returned from the `VmStateIterator` does not contain the operation which was executed to put the VM into this state. We should add another field to this struct so that it looks something like this: ```Rust #[derive(Clone, Debug, Eq, PartialEq)] pub struct VmState { pub op: Operation, pub clk: usize, pub fmp: Felt, pub stack: Vec<Felt>, pub memory: Vec<(u64, Word)>, } ``` The `op` field would need to be populated from the information in the [decoder](https://github.com/maticnetwork/miden/blob/next/processor/src/decoder/mod.rs#L209). The decoder doesn't explicitly track the operations yet. It is possible to infer the operations from the trace, but I think a better approach would be to have a vector of operations in the decoder struct. Then, as the VM executes operations, they would be pushed into this vector. Since tracking operations would result in some overhead, the above should happen only when we are executing programs via [execute_iter()](https://github.com/maticnetwork/miden/blob/next/processor/src/lib.rs#L80) function. Implementing this functionality well enable counting operations executed by the VM which would be useful for things like #198.
process
add operation tracking to vmstateiterator currently the struct which is returned from the vmstateiterator does not contain the operation which was executed to put the vm into this state we should add another field to this struct so that it looks something like this rust pub struct vmstate pub op operation pub clk usize pub fmp felt pub stack vec pub memory vec the op field would need to be populated from the information in the the decoder doesn t explicitly track the operations yet it is possible to infer the operations from the trace but i think a better approach would be to have a vector of operations in the decoder struct then as the vm executes operations they would be pushed into this vector since tracking operations would result in some overhead the above should happen only when we are executing programs via function implementing this functionality well enable counting operations executed by the vm which would be useful for things like
1
259,253
27,621,778,618
IssuesEvent
2023-03-10 01:10:59
praneethpanasala/linux
https://api.github.com/repos/praneethpanasala/linux
opened
CVE-2023-1074 (Medium) detected in linuxlinux-4.19.6
Mend: dependency security vulnerability
## CVE-2023-1074 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.6</b></p></summary> <p> <p>Apache Software Foundation (ASF)</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/sctp/bind_addr.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/sctp/bind_addr.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This CVE was marked as RESERVED by NVD. This means that the CVE-ID is reserved for future use by the CVE Numbering Authority (CNA) or a security researcher. Still, the CVE details were not included in the CVE entry. This is because the original requester of the CVE ID assignment has not sent an update to MITRE with the information needed to populate the CVE entry. This page will reflect the classification results once they are available through NVD <p>Publish Date: 2023-02-28 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-1074>CVE-2023-1074</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-1074">https://www.linuxkernelcves.com/cves/CVE-2023-1074</a></p> <p>Release Date: 2023-02-28</p> <p>Fix Resolution: v4.14.305,v4.19.272,v5.4.231,v5.10.166,v5.15.91,v6.1.9,v6.2-rc6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2023-1074 (Medium) detected in linuxlinux-4.19.6 - ## CVE-2023-1074 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.6</b></p></summary> <p> <p>Apache Software Foundation (ASF)</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/sctp/bind_addr.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/sctp/bind_addr.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This CVE was marked as RESERVED by NVD. This means that the CVE-ID is reserved for future use by the CVE Numbering Authority (CNA) or a security researcher. Still, the CVE details were not included in the CVE entry. This is because the original requester of the CVE ID assignment has not sent an update to MITRE with the information needed to populate the CVE entry. This page will reflect the classification results once they are available through NVD <p>Publish Date: 2023-02-28 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-1074>CVE-2023-1074</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-1074">https://www.linuxkernelcves.com/cves/CVE-2023-1074</a></p> <p>Release Date: 2023-02-28</p> <p>Fix Resolution: v4.14.305,v4.19.272,v5.4.231,v5.10.166,v5.15.91,v6.1.9,v6.2-rc6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux apache software foundation asf library home page a href found in base branch master vulnerable source files net sctp bind addr c net sctp bind addr c vulnerability details this cve was marked as reserved by nvd this means that the cve id is reserved for future use by the cve numbering authority cna or a security researcher still the cve details were not included in the cve entry this is because the original requester of the cve id assignment has not sent an update to mitre with the information needed to populate the cve entry this page will reflect the classification results once they are available through nvd publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
199,916
6,996,123,204
IssuesEvent
2017-12-15 22:30:22
freechipsproject/firrtl
https://api.github.com/repos/freechipsproject/firrtl
closed
Emit Verilog Top-Down
enhancement low priority
So recently, when I was debugging some emitted Verilog, I kept getting tripped up by the fact that logical dependencies kind of went from bottom to top. It doesn't matter as far as synthesizability is concerned, because the signals are all pre-declared, but for some debugging sanity, it'd be super nice if Verilog was emitted top to bottom so the logical flow made more sense.
1.0
Emit Verilog Top-Down - So recently, when I was debugging some emitted Verilog, I kept getting tripped up by the fact that logical dependencies kind of went from bottom to top. It doesn't matter as far as synthesizability is concerned, because the signals are all pre-declared, but for some debugging sanity, it'd be super nice if Verilog was emitted top to bottom so the logical flow made more sense.
non_process
emit verilog top down so recently when i was debugging some emitted verilog i kept getting tripped up by the fact that logical dependencies kind of went from bottom to top it doesn t matter as far as synthesizability is concerned because the signals are all pre declared but for some debugging sanity it d be super nice if verilog was emitted top to bottom so the logical flow made more sense
0
37,749
5,142,132,898
IssuesEvent
2017-01-12 12:15:02
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
opened
ci-kubernetes-e2e-gci-gke-subnet: broken test run
kind/flake priority/P2 team/test-infra
https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-e2e-gci-gke-subnet/191/ Multiple broken tests: Failed: [k8s.io] DNS should provide DNS for the cluster [Conformance] {Kubernetes e2e suite} ``` /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/dns.go:352 Expected error: <*errors.errorString | 0xc42038cc30>: { s: "timed out waiting for the condition", } timed out waiting for the condition not to have occurred /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/dns.go:236 ``` Issues about this test specifically: #26194 #26338 #30345 #34571 Failed: [k8s.io] DNS config map should be able to change configuration {Kubernetes e2e suite} ``` /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/dns_configmap.go:66 Expected error: <*errors.errorString | 0xc42038cc30>: { s: "timed out waiting for the condition", } timed out waiting for the condition not to have occurred /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/dns_configmap.go:283 ``` Issues about this test specifically: #37144 Failed: [k8s.io] DNS should provide DNS for ExternalName services {Kubernetes e2e suite} ``` /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/dns.go:501 Expected error: <*errors.errorString | 0xc42038cc30>: { s: "timed out waiting for the condition", } timed out waiting for the condition not to have occurred /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/dns.go:265 ``` Issues about this test specifically: #32584 Failed: Test {e2e.go} ``` exit status 1 ``` Issues about this test specifically: #33361 #38663 #39788 Previous issues for this suite: #37341 #38356 #38582 #39186
1.0
ci-kubernetes-e2e-gci-gke-subnet: broken test run - https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/ci-kubernetes-e2e-gci-gke-subnet/191/ Multiple broken tests: Failed: [k8s.io] DNS should provide DNS for the cluster [Conformance] {Kubernetes e2e suite} ``` /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/dns.go:352 Expected error: <*errors.errorString | 0xc42038cc30>: { s: "timed out waiting for the condition", } timed out waiting for the condition not to have occurred /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/dns.go:236 ``` Issues about this test specifically: #26194 #26338 #30345 #34571 Failed: [k8s.io] DNS config map should be able to change configuration {Kubernetes e2e suite} ``` /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/dns_configmap.go:66 Expected error: <*errors.errorString | 0xc42038cc30>: { s: "timed out waiting for the condition", } timed out waiting for the condition not to have occurred /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/dns_configmap.go:283 ``` Issues about this test specifically: #37144 Failed: [k8s.io] DNS should provide DNS for ExternalName services {Kubernetes e2e suite} ``` /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/dns.go:501 Expected error: <*errors.errorString | 0xc42038cc30>: { s: "timed out waiting for the condition", } timed out waiting for the condition not to have occurred /go/src/k8s.io/kubernetes/_output/dockerized/go/src/k8s.io/kubernetes/test/e2e/dns.go:265 ``` Issues about this test specifically: #32584 Failed: Test {e2e.go} ``` exit status 1 ``` Issues about this test specifically: #33361 #38663 #39788 Previous issues for this suite: #37341 #38356 #38582 #39186
non_process
ci kubernetes gci gke subnet broken test run multiple broken tests failed dns should provide dns for the cluster kubernetes suite go src io kubernetes output dockerized go src io kubernetes test dns go expected error s timed out waiting for the condition timed out waiting for the condition not to have occurred go src io kubernetes output dockerized go src io kubernetes test dns go issues about this test specifically failed dns config map should be able to change configuration kubernetes suite go src io kubernetes output dockerized go src io kubernetes test dns configmap go expected error s timed out waiting for the condition timed out waiting for the condition not to have occurred go src io kubernetes output dockerized go src io kubernetes test dns configmap go issues about this test specifically failed dns should provide dns for externalname services kubernetes suite go src io kubernetes output dockerized go src io kubernetes test dns go expected error s timed out waiting for the condition timed out waiting for the condition not to have occurred go src io kubernetes output dockerized go src io kubernetes test dns go issues about this test specifically failed test go exit status issues about this test specifically previous issues for this suite
0
4,623
7,468,690,066
IssuesEvent
2018-04-02 19:54:44
w3c/w3process
https://api.github.com/repos/w3c/w3process
closed
Process2014 introduced an AC ballot for CR transitions and now we have a ballot open for a year
Process2018Candidate
Transferred from https://www.w3.org/community/w3process/track/issues/182 State: Raised
1.0
Process2014 introduced an AC ballot for CR transitions and now we have a ballot open for a year - Transferred from https://www.w3.org/community/w3process/track/issues/182 State: Raised
process
introduced an ac ballot for cr transitions and now we have a ballot open for a year transferred from state raised
1
12,359
14,888,287,582
IssuesEvent
2021-01-20 19:37:06
CodeForPittsburgh/food-access-map-data
https://api.github.com/repos/CodeForPittsburgh/food-access-map-data
closed
process new farmers market data
data processing
the new farmers market datasource needs to be wrangled into our schema.
1.0
process new farmers market data - the new farmers market datasource needs to be wrangled into our schema.
process
process new farmers market data the new farmers market datasource needs to be wrangled into our schema
1
61,118
6,725,781,409
IssuesEvent
2017-10-17 07:25:35
openshift/origin
https://api.github.com/repos/openshift/origin
closed
flake: templateservicebroker security test should pass security tests
area/security kind/test-flake priority/P0
https://openshift-gce-devel.appspot.com/build/origin-ci-test/pr-logs/pull/16851/test_pull_request_origin_extended_conformance_gce/9755/ ``` [Conformance][templates] templateservicebroker security test should pass security tests [Suite:openshift/conformance/parallel] 1m43s go run hack/e2e.go -v -test --test_args='--ginkgo.focus=\s\[Conformance\]\[templates\]\stemplateservicebroker\ssecurity\stest\sshould\spass\ssecurity\stests\s\[Suite\:openshift\/conformance\/parallel\]$' /tmp/openshift/build-rpm-release/tito/rpmbuild-origin97017y/BUILD/origin-3.7.0/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateservicebroker_security.go:257 Expected error: <*client.ServerError | 0xc421d8b3c0>: { StatusCode: 403, Description: "secrets \"d2a2b8b4-b577-42c8-92b7-0f5b85d03592\" is forbidden: cannot set blockOwnerDeletion if an ownerReference refers to a resource you can't set finalizers on: User \"system:serviceaccount:extended-test-openshift-template-service-broker-j0pr1-hp4dk:apiserver\" cannot update brokertemplateinstances/finalizers.template.openshift.io in project \"extended-test-templates-46xbx-whdrr\", <nil>", } Forbidden: secrets "d2a2b8b4-b577-42c8-92b7-0f5b85d03592" is forbidden: cannot set blockOwnerDeletion if an ownerReference refers to a resource you can't set finalizers on: User "system:serviceaccount:extended-test-openshift-template-service-broker-j0pr1-hp4dk:apiserver" cannot update brokertemplateinstances/finalizers.template.openshift.io in project "extended-test-templates-46xbx-whdrr", <nil> not to have occurred /tmp/openshift/build-rpm-release/tito/rpmbuild-origin97017y/BUILD/origin-3.7.0/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateservicebroker_security.go:162 ``` Looks like permission cache not being filled yet? @openshift/sig-security @openshift/sig-master
1.0
flake: templateservicebroker security test should pass security tests - https://openshift-gce-devel.appspot.com/build/origin-ci-test/pr-logs/pull/16851/test_pull_request_origin_extended_conformance_gce/9755/ ``` [Conformance][templates] templateservicebroker security test should pass security tests [Suite:openshift/conformance/parallel] 1m43s go run hack/e2e.go -v -test --test_args='--ginkgo.focus=\s\[Conformance\]\[templates\]\stemplateservicebroker\ssecurity\stest\sshould\spass\ssecurity\stests\s\[Suite\:openshift\/conformance\/parallel\]$' /tmp/openshift/build-rpm-release/tito/rpmbuild-origin97017y/BUILD/origin-3.7.0/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateservicebroker_security.go:257 Expected error: <*client.ServerError | 0xc421d8b3c0>: { StatusCode: 403, Description: "secrets \"d2a2b8b4-b577-42c8-92b7-0f5b85d03592\" is forbidden: cannot set blockOwnerDeletion if an ownerReference refers to a resource you can't set finalizers on: User \"system:serviceaccount:extended-test-openshift-template-service-broker-j0pr1-hp4dk:apiserver\" cannot update brokertemplateinstances/finalizers.template.openshift.io in project \"extended-test-templates-46xbx-whdrr\", <nil>", } Forbidden: secrets "d2a2b8b4-b577-42c8-92b7-0f5b85d03592" is forbidden: cannot set blockOwnerDeletion if an ownerReference refers to a resource you can't set finalizers on: User "system:serviceaccount:extended-test-openshift-template-service-broker-j0pr1-hp4dk:apiserver" cannot update brokertemplateinstances/finalizers.template.openshift.io in project "extended-test-templates-46xbx-whdrr", <nil> not to have occurred /tmp/openshift/build-rpm-release/tito/rpmbuild-origin97017y/BUILD/origin-3.7.0/_output/local/go/src/github.com/openshift/origin/test/extended/templates/templateservicebroker_security.go:162 ``` Looks like permission cache not being filled yet? @openshift/sig-security @openshift/sig-master
non_process
flake templateservicebroker security test should pass security tests templateservicebroker security test should pass security tests go run hack go v test test args ginkgo focus s stemplateservicebroker ssecurity stest sshould spass ssecurity stests s tmp openshift build rpm release tito rpmbuild build origin output local go src github com openshift origin test extended templates templateservicebroker security go expected error statuscode description secrets is forbidden cannot set blockownerdeletion if an ownerreference refers to a resource you can t set finalizers on user system serviceaccount extended test openshift template service broker apiserver cannot update brokertemplateinstances finalizers template openshift io in project extended test templates whdrr forbidden secrets is forbidden cannot set blockownerdeletion if an ownerreference refers to a resource you can t set finalizers on user system serviceaccount extended test openshift template service broker apiserver cannot update brokertemplateinstances finalizers template openshift io in project extended test templates whdrr not to have occurred tmp openshift build rpm release tito rpmbuild build origin output local go src github com openshift origin test extended templates templateservicebroker security go looks like permission cache not being filled yet openshift sig security openshift sig master
0
17,622
23,442,524,939
IssuesEvent
2022-08-15 16:13:24
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
Can't pickle local object 'CDLL.__init__.<locals>._FuncPtr'
module: multiprocessing triaged
I am using `torch.multiprocessing.Process` in a reinforcement learning project. Here's bit of my codes: ```python class Agent(mp.Process): def __init__(): super()... self.dll = CDLL('load a dll here which work fine without mp.Process') self.env = some_env def run(self): action = ... self.env.step(action, self.dll) ... if __name__ == "__main__": mp.set_start_method("forkserver") workers = [Agent(..) for i in range(n)] [w.start() for w in workers] [w.join() for w in workers] ``` Here's the errors I got: ``` Traceback (most recent call last): File "gym_test.py", line 173, in <module> [w.start() for w in workers] File "gym_test.py", line 173, in <listcomp> [w.start() for w in workers] File "/usr/lib/python3.7/multiprocessing/process.py", line 112, in start self._popen = self._Popen(self) File "/usr/lib/python3.7/multiprocessing/context.py", line 223, in _Popen return _default_context.get_context().Process._Popen(process_obj) File "/usr/lib/python3.7/multiprocessing/context.py", line 291, in _Popen return Popen(process_obj) File "/usr/lib/python3.7/multiprocessing/popen_forkserver.py", line 35, in __init__ super().__init__(process_obj) File "/usr/lib/python3.7/multiprocessing/popen_fork.py", line 20, in __init__ self._launch(process_obj) File "/usr/lib/python3.7/multiprocessing/popen_forkserver.py", line 47, in _launch reduction.dump(process_obj, buf) File "/usr/lib/python3.7/multiprocessing/reduction.py", line 60, in dump ForkingPickler(file, protocol).dump(obj) AttributeError: Can't pickle local object 'CDLL.__init__.<locals>._FuncPtr' ``` I wonder how to load cdll properly in multiprocessing, any solutions? cc @VitalyFedyunin
1.0
Can't pickle local object 'CDLL.__init__.<locals>._FuncPtr' - I am using `torch.multiprocessing.Process` in a reinforcement learning project. Here's bit of my codes: ```python class Agent(mp.Process): def __init__(): super()... self.dll = CDLL('load a dll here which work fine without mp.Process') self.env = some_env def run(self): action = ... self.env.step(action, self.dll) ... if __name__ == "__main__": mp.set_start_method("forkserver") workers = [Agent(..) for i in range(n)] [w.start() for w in workers] [w.join() for w in workers] ``` Here's the errors I got: ``` Traceback (most recent call last): File "gym_test.py", line 173, in <module> [w.start() for w in workers] File "gym_test.py", line 173, in <listcomp> [w.start() for w in workers] File "/usr/lib/python3.7/multiprocessing/process.py", line 112, in start self._popen = self._Popen(self) File "/usr/lib/python3.7/multiprocessing/context.py", line 223, in _Popen return _default_context.get_context().Process._Popen(process_obj) File "/usr/lib/python3.7/multiprocessing/context.py", line 291, in _Popen return Popen(process_obj) File "/usr/lib/python3.7/multiprocessing/popen_forkserver.py", line 35, in __init__ super().__init__(process_obj) File "/usr/lib/python3.7/multiprocessing/popen_fork.py", line 20, in __init__ self._launch(process_obj) File "/usr/lib/python3.7/multiprocessing/popen_forkserver.py", line 47, in _launch reduction.dump(process_obj, buf) File "/usr/lib/python3.7/multiprocessing/reduction.py", line 60, in dump ForkingPickler(file, protocol).dump(obj) AttributeError: Can't pickle local object 'CDLL.__init__.<locals>._FuncPtr' ``` I wonder how to load cdll properly in multiprocessing, any solutions? cc @VitalyFedyunin
process
can t pickle local object cdll init funcptr i am using torch multiprocessing process in a reinforcement learning project here s bit of my codes python class agent mp process def init super self dll cdll load a dll here which work fine without mp process self env some env def run self action self env step action self dll if name main mp set start method forkserver workers here s the errors i got traceback most recent call last file gym test py line in file gym test py line in file usr lib multiprocessing process py line in start self popen self popen self file usr lib multiprocessing context py line in popen return default context get context process popen process obj file usr lib multiprocessing context py line in popen return popen process obj file usr lib multiprocessing popen forkserver py line in init super init process obj file usr lib multiprocessing popen fork py line in init self launch process obj file usr lib multiprocessing popen forkserver py line in launch reduction dump process obj buf file usr lib multiprocessing reduction py line in dump forkingpickler file protocol dump obj attributeerror can t pickle local object cdll init funcptr i wonder how to load cdll properly in multiprocessing any solutions cc vitalyfedyunin
1
50,373
6,367,231,791
IssuesEvent
2017-08-01 05:20:58
openMF/community-app
https://api.github.com/repos/openMF/community-app
closed
Charge Fee Frequency is not displaying properly.
bug design gsoc p1
1. Go to Admin>Products>Charges and Create Charge 2. Select Loan for "Charge Applies To" 3. Furnish the inputs and select 'Charge Time Type' as Overdue Fees 4. Enable 'Add Frequency' and click on drop-down menu of Charge Frequency. >Check the following Screenshot: ![overduefee](https://user-images.githubusercontent.com/23022667/28764618-0b02af14-75e4-11e7-95e5-99d36aebe485.png) Expected: It should just display Days, Weeks, Months and Years.
1.0
Charge Fee Frequency is not displaying properly. - 1. Go to Admin>Products>Charges and Create Charge 2. Select Loan for "Charge Applies To" 3. Furnish the inputs and select 'Charge Time Type' as Overdue Fees 4. Enable 'Add Frequency' and click on drop-down menu of Charge Frequency. >Check the following Screenshot: ![overduefee](https://user-images.githubusercontent.com/23022667/28764618-0b02af14-75e4-11e7-95e5-99d36aebe485.png) Expected: It should just display Days, Weeks, Months and Years.
non_process
charge fee frequency is not displaying properly go to admin products charges and create charge select loan for charge applies to furnish the inputs and select charge time type as overdue fees enable add frequency and click on drop down menu of charge frequency check the following screenshot expected it should just display days weeks months and years
0
795,891
28,091,058,910
IssuesEvent
2023-03-30 13:05:42
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
twitch.tv - site is not usable
browser-firefox priority-important os-linux engine-gecko
<!-- @browser: Firefox 111.0 --> <!-- @ua_header: Mozilla/5.0 (X11; Linux x86_64; rv:109.0) Gecko/20100101 Firefox/111.0 --> <!-- @reported_with: unknown --> **URL**: https://twitch.tv **Browser / Version**: Firefox 111.0 **Operating System**: Linux **Tested Another Browser**: Yes Chrome **Problem type**: Site is not usable **Description**: Browser unsupported **Steps to Reproduce**: Login page says browser is unsupported <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2023/3/22c79761-d36e-4de1-911a-f6f4bba1b27e.jpg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
twitch.tv - site is not usable - <!-- @browser: Firefox 111.0 --> <!-- @ua_header: Mozilla/5.0 (X11; Linux x86_64; rv:109.0) Gecko/20100101 Firefox/111.0 --> <!-- @reported_with: unknown --> **URL**: https://twitch.tv **Browser / Version**: Firefox 111.0 **Operating System**: Linux **Tested Another Browser**: Yes Chrome **Problem type**: Site is not usable **Description**: Browser unsupported **Steps to Reproduce**: Login page says browser is unsupported <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2023/3/22c79761-d36e-4de1-911a-f6f4bba1b27e.jpg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
twitch tv site is not usable url browser version firefox operating system linux tested another browser yes chrome problem type site is not usable description browser unsupported steps to reproduce login page says browser is unsupported view the screenshot img alt screenshot src browser configuration none from with ❤️
0
6,622
9,725,389,829
IssuesEvent
2019-05-30 08:34:12
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Scan data for attributes to use in a Processing model
Feature Request Processing
Author Name: **Magnus Nilsson** (Magnus Nilsson) Original Redmine Issue: [20986](https://issues.qgis.org/issues/20986) Redmine category:processing/modeller --- I am not too fond of the idea of manually having to enter which attributes to use in a Processing model. For example, in the tool for creating points from a table, I manually have to enter (write) which fields to use for x/y coordinates. Wouldn´t it be better to allow a model to scan the data for attributes and let the user select relevant attributes from a list? Related: When adding a vector layer to a model, why not choose the data source directly? That way, the model knows the available attributes. Having to define this each time I run a model is time consuming.
1.0
Scan data for attributes to use in a Processing model - Author Name: **Magnus Nilsson** (Magnus Nilsson) Original Redmine Issue: [20986](https://issues.qgis.org/issues/20986) Redmine category:processing/modeller --- I am not too fond of the idea of manually having to enter which attributes to use in a Processing model. For example, in the tool for creating points from a table, I manually have to enter (write) which fields to use for x/y coordinates. Wouldn´t it be better to allow a model to scan the data for attributes and let the user select relevant attributes from a list? Related: When adding a vector layer to a model, why not choose the data source directly? That way, the model knows the available attributes. Having to define this each time I run a model is time consuming.
process
scan data for attributes to use in a processing model author name magnus nilsson magnus nilsson original redmine issue redmine category processing modeller i am not too fond of the idea of manually having to enter which attributes to use in a processing model for example in the tool for creating points from a table i manually have to enter write which fields to use for x y coordinates wouldn´t it be better to allow a model to scan the data for attributes and let the user select relevant attributes from a list related when adding a vector layer to a model why not choose the data source directly that way the model knows the available attributes having to define this each time i run a model is time consuming
1
131,113
27,824,457,756
IssuesEvent
2023-03-19 15:58:25
pinterest/ktlint
https://api.github.com/repos/pinterest/ktlint
closed
Indentation change since 0.38.1
indentation-rule conflict-with-default-intellij-formatting ktlint-official-codestyle
## Expected Behavior Upgrading from 0.38.1 to latest I see a change in indentation with parameter names in a call. For example this was allowed in 0.38.1: someFunction( parameterName = someValue .someProperty .someCall() ) ## Observed Behavior With 0.41.1 it now demands that it be indented like this: someFunction( parameterName = someValue .someProperty .someCall() ) Which is not very readable. I can make it go away by moving someValue to same line as parameter name: someFunction( parameterName = someValue .someProperty .someCall() ) But sometimes there is a longer expression than this example and it is more readable to move it to its own line.
1.0
Indentation change since 0.38.1 - ## Expected Behavior Upgrading from 0.38.1 to latest I see a change in indentation with parameter names in a call. For example this was allowed in 0.38.1: someFunction( parameterName = someValue .someProperty .someCall() ) ## Observed Behavior With 0.41.1 it now demands that it be indented like this: someFunction( parameterName = someValue .someProperty .someCall() ) Which is not very readable. I can make it go away by moving someValue to same line as parameter name: someFunction( parameterName = someValue .someProperty .someCall() ) But sometimes there is a longer expression than this example and it is more readable to move it to its own line.
non_process
indentation change since expected behavior upgrading from to latest i see a change in indentation with parameter names in a call for example this was allowed in somefunction parametername somevalue someproperty somecall observed behavior with it now demands that it be indented like this somefunction parametername somevalue someproperty somecall which is not very readable i can make it go away by moving somevalue to same line as parameter name somefunction parametername somevalue someproperty somecall but sometimes there is a longer expression than this example and it is more readable to move it to its own line
0
3,331
6,447,850,646
IssuesEvent
2017-08-14 09:22:41
nodejs/node
https://api.github.com/repos/nodejs/node
closed
maxBuffer default too small
child_process
* **Version**: v6.2.1 * **Platform**: Darwin me.local 15.5.0 Darwin Kernel Version 15.5.0: Tue Apr 19 18:36:36 PDT 2016; root:xnu-3248.50.21~8/RELEASE_X86_64 x86_64 * **Subsystem**: child_process Currently `maxBuffer` for `child_process.exec` is set to 200*1024 bytes, or ~204.8KB. I ran into an issue where my child process was being terminated and tracking it down was quite tough. It ended up being that it was producing enough output that it exceeded `maxBuffer`. I think the buffer size is too small and this behavior (terminating a child) is drastic enough that it should only be done in the case where a child is producing a much larger amount of output. I'm not sure what's sane here, perhaps 5MB+?
1.0
maxBuffer default too small - * **Version**: v6.2.1 * **Platform**: Darwin me.local 15.5.0 Darwin Kernel Version 15.5.0: Tue Apr 19 18:36:36 PDT 2016; root:xnu-3248.50.21~8/RELEASE_X86_64 x86_64 * **Subsystem**: child_process Currently `maxBuffer` for `child_process.exec` is set to 200*1024 bytes, or ~204.8KB. I ran into an issue where my child process was being terminated and tracking it down was quite tough. It ended up being that it was producing enough output that it exceeded `maxBuffer`. I think the buffer size is too small and this behavior (terminating a child) is drastic enough that it should only be done in the case where a child is producing a much larger amount of output. I'm not sure what's sane here, perhaps 5MB+?
process
maxbuffer default too small version platform darwin me local darwin kernel version tue apr pdt root xnu release subsystem child process currently maxbuffer for child process exec is set to bytes or i ran into an issue where my child process was being terminated and tracking it down was quite tough it ended up being that it was producing enough output that it exceeded maxbuffer i think the buffer size is too small and this behavior terminating a child is drastic enough that it should only be done in the case where a child is producing a much larger amount of output i m not sure what s sane here perhaps
1
728,712
25,089,164,885
IssuesEvent
2022-11-08 03:53:08
ballerina-platform/ballerina-dev-website
https://api.github.com/repos/ballerina-platform/ballerina-dev-website
closed
Improve styles in newsletter
Priority/Highest Type/Improvement
## Description > $subject ## Related website/documentation area > Add/Uncomment the relevant area label out of the following. <!--Area/BBEs--> <!--Area/HomePageSamples--> <!--Area/LearnPages--> <!--Area/CommonPages--> <!--Area/Backend--> <!--Area/UIUX--> <!--Area/Workflows--> <!--Area/Blog--> ## Describe the problem(s) > A detailed description of the purpose of this improvement. ## Describe your solution(s) > A detailed description of how this improvement will solve the problem described above. ## Related issue(s) (optional) > Any related issues such as sub tasks and issues reported in other repositories (e.g., component repositories), similar problems, etc. ## Suggested label(s) (optional) > Optional comma-separated list of suggested labels. Non committers can’t assign labels to issues, and thereby, this will help issue creators who are not a committer to suggest possible labels. ## Suggested assignee(s) (optional) > Optional comma-separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, and thereby, this will help issue creators who are not a committer to suggest possible assignees.
1.0
Improve styles in newsletter - ## Description > $subject ## Related website/documentation area > Add/Uncomment the relevant area label out of the following. <!--Area/BBEs--> <!--Area/HomePageSamples--> <!--Area/LearnPages--> <!--Area/CommonPages--> <!--Area/Backend--> <!--Area/UIUX--> <!--Area/Workflows--> <!--Area/Blog--> ## Describe the problem(s) > A detailed description of the purpose of this improvement. ## Describe your solution(s) > A detailed description of how this improvement will solve the problem described above. ## Related issue(s) (optional) > Any related issues such as sub tasks and issues reported in other repositories (e.g., component repositories), similar problems, etc. ## Suggested label(s) (optional) > Optional comma-separated list of suggested labels. Non committers can’t assign labels to issues, and thereby, this will help issue creators who are not a committer to suggest possible labels. ## Suggested assignee(s) (optional) > Optional comma-separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, and thereby, this will help issue creators who are not a committer to suggest possible assignees.
non_process
improve styles in newsletter description subject related website documentation area add uncomment the relevant area label out of the following describe the problem s a detailed description of the purpose of this improvement describe your solution s a detailed description of how this improvement will solve the problem described above related issue s optional any related issues such as sub tasks and issues reported in other repositories e g component repositories similar problems etc suggested label s optional optional comma separated list of suggested labels non committers can’t assign labels to issues and thereby this will help issue creators who are not a committer to suggest possible labels suggested assignee s optional optional comma separated list of suggested team members who should attend the issue non committers can’t assign issues to assignees and thereby this will help issue creators who are not a committer to suggest possible assignees
0
448,689
31,809,004,675
IssuesEvent
2023-09-13 15:34:20
vijayk3327/Visualforce
https://api.github.com/repos/vijayk3327/Visualforce
opened
How to Get Billing Address of Account Object from Quote based on record Id using Apex Class Method in Visualforce Page Salesforce
documentation question
In this post we are going to learn about How to Get Billing Address of Account Object from Quote based on record Id using Apex Class Method in Visualforce Page Salesforce. Address is a “compound field” – in simple terms, this means the field groups together multiple individual elements into a single compound. Custom Address Fields:- Users can populate a custom address fields manually or they can use the Google lookup to search for an address. Admins and APIs can access each address stored in a custom address field as a structured compound data type as well as individual address components. **[👉 Get source code live demo link:-](https://www.w3web.net/get-account-billing-address-from-quote/)** <img src="https://www.w3web.net/wp-content/uploads/2023/05/accAddressFromQuote-min.gif"/> `Create Visualforce Page Step 1:- Create Visualforce Page : accAddressFromQuoteVf.vfp` `<apex:page standardController="Quote" extensions="accAddressFromQuoteVfCtrl"> <div class="slds slds-p-around_small" style="padding:10px;"> <table width="50%" border="1" cellspacing="0" cellpadding="5" bordercolor="#ccc" class="slds-table slds-table_bordered slds-table_col-bordered" style="border-collapse:collapse;"> <tr style="background:#ddd;"> <th>Quote Name</th> <th>Billing Street</th> <th>Billing City</th> <th>Billing Country</th> <th>Billing Postal Code</th> </tr> <apex:repeat value="{!quoteObj}" var="qotItem"> <tr> <td><p><apex:outputText value="{!qotItem.Name}"/></p></td> <td><p><apex:outputText value="{!qotItem.Account.BillingStreet}"/></p></td> <td><p><apex:outputText value="{!qotItem.Account.BillingCity}"/></p></td> <td><p><apex:outputText value="{!qotItem.Account.BillingCountry}"/></p></td> <td><p><apex:outputText value="{!qotItem.Account.BillingPostalCode}"/></p></td> </tr> </apex:repeat> </table> </div> </apex:page>` **Create Apex Class Extension Controller in Visualforce Step 2:- Create Apex Class : accAddressFromQuoteVfCtrl.apxc** `public class accAddressFromQuoteVfCtrl { public String MstrId{GET;SET;} public Quote quoteObj{GET;SET;} public accAddressFromQuoteVfCtrl(ApexPages.StandardController Controller){ MstrId = ApexPages.currentPage().getParameters().get('id'); quoteObj = [SELECT Id, Name, AccountId, Account.Name, Account.BillingStreet, Account.BillingCity, Account.BillingCountry, Account.BillingPostalCode FROM Quote WHERE Id=:MstrId ]; } }` **[👉 Get source code live demo link:-](https://www.w3web.net/get-account-billing-address-from-quote/)**
1.0
How to Get Billing Address of Account Object from Quote based on record Id using Apex Class Method in Visualforce Page Salesforce - In this post we are going to learn about How to Get Billing Address of Account Object from Quote based on record Id using Apex Class Method in Visualforce Page Salesforce. Address is a “compound field” – in simple terms, this means the field groups together multiple individual elements into a single compound. Custom Address Fields:- Users can populate a custom address fields manually or they can use the Google lookup to search for an address. Admins and APIs can access each address stored in a custom address field as a structured compound data type as well as individual address components. **[👉 Get source code live demo link:-](https://www.w3web.net/get-account-billing-address-from-quote/)** <img src="https://www.w3web.net/wp-content/uploads/2023/05/accAddressFromQuote-min.gif"/> `Create Visualforce Page Step 1:- Create Visualforce Page : accAddressFromQuoteVf.vfp` `<apex:page standardController="Quote" extensions="accAddressFromQuoteVfCtrl"> <div class="slds slds-p-around_small" style="padding:10px;"> <table width="50%" border="1" cellspacing="0" cellpadding="5" bordercolor="#ccc" class="slds-table slds-table_bordered slds-table_col-bordered" style="border-collapse:collapse;"> <tr style="background:#ddd;"> <th>Quote Name</th> <th>Billing Street</th> <th>Billing City</th> <th>Billing Country</th> <th>Billing Postal Code</th> </tr> <apex:repeat value="{!quoteObj}" var="qotItem"> <tr> <td><p><apex:outputText value="{!qotItem.Name}"/></p></td> <td><p><apex:outputText value="{!qotItem.Account.BillingStreet}"/></p></td> <td><p><apex:outputText value="{!qotItem.Account.BillingCity}"/></p></td> <td><p><apex:outputText value="{!qotItem.Account.BillingCountry}"/></p></td> <td><p><apex:outputText value="{!qotItem.Account.BillingPostalCode}"/></p></td> </tr> </apex:repeat> </table> </div> </apex:page>` **Create Apex Class Extension Controller in Visualforce Step 2:- Create Apex Class : accAddressFromQuoteVfCtrl.apxc** `public class accAddressFromQuoteVfCtrl { public String MstrId{GET;SET;} public Quote quoteObj{GET;SET;} public accAddressFromQuoteVfCtrl(ApexPages.StandardController Controller){ MstrId = ApexPages.currentPage().getParameters().get('id'); quoteObj = [SELECT Id, Name, AccountId, Account.Name, Account.BillingStreet, Account.BillingCity, Account.BillingCountry, Account.BillingPostalCode FROM Quote WHERE Id=:MstrId ]; } }` **[👉 Get source code live demo link:-](https://www.w3web.net/get-account-billing-address-from-quote/)**
non_process
how to get billing address of account object from quote based on record id using apex class method in visualforce page salesforce in this post we are going to learn about how to get billing address of account object from quote based on record id using apex class method in visualforce page salesforce address is a “compound field” – in simple terms this means the field groups together multiple individual elements into a single compound custom address fields users can populate a custom address fields manually or they can use the google lookup to search for an address admins and apis can access each address stored in a custom address field as a structured compound data type as well as individual address components img src create visualforce page step create visualforce page accaddressfromquotevf vfp quote name billing street billing city billing country billing postal code create apex class extension controller in visualforce step create apex class accaddressfromquotevfctrl apxc public class accaddressfromquotevfctrl public string mstrid get set public quote quoteobj get set public accaddressfromquotevfctrl apexpages standardcontroller controller mstrid apexpages currentpage getparameters get id quoteobj
0
2,314
5,135,355,155
IssuesEvent
2017-01-11 12:03:29
jlm2017/jlm-video-subtitles
https://api.github.com/repos/jlm2017/jlm-video-subtitles
opened
[Subtitles] [FR] MÉLENCHON - Déboulé à Tourcoing
Language: French Process: [0] Awaiting subtitles
# **Video title** MÉLENCHON - Déboulé à Tourcoing # **Url** https://www.youtube.com/watch?v=NNe0bXIJ9cQ&t=593s # **Youtube subtitles language** Français # **Duration** 1:41:42 # **Subtitles URL** https://www.youtube.com/timedtext_editor?ref=player&action_mde_edit_form=1&lang=fr&tab=captions&ui=hd&v=NNe0bXIJ9cQ&bl=vmp
1.0
[Subtitles] [FR] MÉLENCHON - Déboulé à Tourcoing - # **Video title** MÉLENCHON - Déboulé à Tourcoing # **Url** https://www.youtube.com/watch?v=NNe0bXIJ9cQ&t=593s # **Youtube subtitles language** Français # **Duration** 1:41:42 # **Subtitles URL** https://www.youtube.com/timedtext_editor?ref=player&action_mde_edit_form=1&lang=fr&tab=captions&ui=hd&v=NNe0bXIJ9cQ&bl=vmp
process
mélenchon déboulé à tourcoing video title mélenchon déboulé à tourcoing url youtube subtitles language français duration subtitles url
1
350,647
31,931,880,542
IssuesEvent
2023-09-19 08:00:07
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
reopened
Fix linalg.test_tensorflow_inv
TensorFlow Frontend Sub Task Failing Test
| | | |---|---| |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/6005627299/job/16288695179"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/6005627299/job/16288695179"><img src=https://img.shields.io/badge/-failure-red></a> |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/6005627299/job/16288695179"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/6005627299/job/16288695179"><img src=https://img.shields.io/badge/-failure-red></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/6005627299/job/16288695179"><img src=https://img.shields.io/badge/-failure-red></a>
1.0
Fix linalg.test_tensorflow_inv - | | | |---|---| |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/6005627299/job/16288695179"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/6005627299/job/16288695179"><img src=https://img.shields.io/badge/-failure-red></a> |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/6005627299/job/16288695179"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/6005627299/job/16288695179"><img src=https://img.shields.io/badge/-failure-red></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/6005627299/job/16288695179"><img src=https://img.shields.io/badge/-failure-red></a>
non_process
fix linalg test tensorflow inv numpy a href src jax a href src tensorflow a href src torch a href src paddle a href src
0
229,201
18,286,657,022
IssuesEvent
2021-10-05 11:04:03
DILCISBoard/eark-ip-test-corpus
https://api.github.com/repos/DILCISBoard/eark-ip-test-corpus
closed
CSIP42 Test Case Description
test case corpus package
**Specification:** - **Name:** E-ARK CSIP - **Version:** 2.0-DRAFT - **URL:** http://earkcsip.dilcis.eu/ **Requirement:** - **Id:** CSIP42 - **Link:** http://earkcsip.dilcis.eu/#CSIP42 **Error Level:** ERROR **Description:** CSIP42 | File creation date amdSec/digiprovMD/mdRef/@CREATED | Date the linked file was created. | 1..1 MUST -- | -- | -- | --
1.0
CSIP42 Test Case Description - **Specification:** - **Name:** E-ARK CSIP - **Version:** 2.0-DRAFT - **URL:** http://earkcsip.dilcis.eu/ **Requirement:** - **Id:** CSIP42 - **Link:** http://earkcsip.dilcis.eu/#CSIP42 **Error Level:** ERROR **Description:** CSIP42 | File creation date amdSec/digiprovMD/mdRef/@CREATED | Date the linked file was created. | 1..1 MUST -- | -- | -- | --
non_process
test case description specification name e ark csip version draft url requirement id link error level error description file creation date amdsec digiprovmd mdref created date the linked file was created must
0
260,276
27,781,305,315
IssuesEvent
2023-03-16 21:19:39
snowdensb/caseflow
https://api.github.com/repos/snowdensb/caseflow
opened
CVE-2023-28155 (Medium) detected in request-2.88.2.tgz
Mend: dependency security vulnerability
## CVE-2023-28155 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>request-2.88.2.tgz</b></p></summary> <p>Simplified HTTP request client.</p> <p>Library home page: <a href="https://registry.npmjs.org/request/-/request-2.88.2.tgz">https://registry.npmjs.org/request/-/request-2.88.2.tgz</a></p> <p>Path to dependency file: /client/package.json</p> <p>Path to vulnerable library: /client/node_modules/request/package.json</p> <p> Dependency Hierarchy: - jsdom-9.8.3.tgz (Root Library) - :x: **request-2.88.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/snowdensb/caseflow/commit/81f8b3f5658022f994993a18a7653667705b7f6e">81f8b3f5658022f994993a18a7653667705b7f6e</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ** UNSUPPORTED WHEN ASSIGNED ** The Request package through 2.88.1 for Node.js allows a bypass of SSRF mitigations via an attacker-controller server that does a cross-protocol redirect (HTTP to HTTPS, or HTTPS to HTTP). NOTE: This vulnerability only affects products that are no longer supported by the maintainer. <p>Publish Date: 2023-03-16 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-28155>CVE-2023-28155</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p>
True
CVE-2023-28155 (Medium) detected in request-2.88.2.tgz - ## CVE-2023-28155 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>request-2.88.2.tgz</b></p></summary> <p>Simplified HTTP request client.</p> <p>Library home page: <a href="https://registry.npmjs.org/request/-/request-2.88.2.tgz">https://registry.npmjs.org/request/-/request-2.88.2.tgz</a></p> <p>Path to dependency file: /client/package.json</p> <p>Path to vulnerable library: /client/node_modules/request/package.json</p> <p> Dependency Hierarchy: - jsdom-9.8.3.tgz (Root Library) - :x: **request-2.88.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/snowdensb/caseflow/commit/81f8b3f5658022f994993a18a7653667705b7f6e">81f8b3f5658022f994993a18a7653667705b7f6e</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ** UNSUPPORTED WHEN ASSIGNED ** The Request package through 2.88.1 for Node.js allows a bypass of SSRF mitigations via an attacker-controller server that does a cross-protocol redirect (HTTP to HTTPS, or HTTPS to HTTP). NOTE: This vulnerability only affects products that are no longer supported by the maintainer. <p>Publish Date: 2023-03-16 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-28155>CVE-2023-28155</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p>
non_process
cve medium detected in request tgz cve medium severity vulnerability vulnerable library request tgz simplified http request client library home page a href path to dependency file client package json path to vulnerable library client node modules request package json dependency hierarchy jsdom tgz root library x request tgz vulnerable library found in head commit a href found in base branch master vulnerability details unsupported when assigned the request package through for node js allows a bypass of ssrf mitigations via an attacker controller server that does a cross protocol redirect http to https or https to http note this vulnerability only affects products that are no longer supported by the maintainer publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href
0
1,775
2,666,940,981
IssuesEvent
2015-03-22 02:35:30
benquarmby/jslintnet-test
https://api.github.com/repos/benquarmby/jslintnet-test
opened
Ability to upgrade JSLint version without recompiling
CodePlex
<b>ChrisNielsen[CodePlex]</b> <br />I am using the MSBuild task. I would like to be able to upgrade the JSLint version without needing to recompile JSLintNet. If it could read jslint.js from the file system instead of as an embedded resource, that would be great.
1.0
Ability to upgrade JSLint version without recompiling - <b>ChrisNielsen[CodePlex]</b> <br />I am using the MSBuild task. I would like to be able to upgrade the JSLint version without needing to recompile JSLintNet. If it could read jslint.js from the file system instead of as an embedded resource, that would be great.
non_process
ability to upgrade jslint version without recompiling chrisnielsen i am using the msbuild task i would like to be able to upgrade the jslint version without needing to recompile jslintnet if it could read jslint js from the file system instead of as an embedded resource that would be great
0
10,646
13,446,217,972
IssuesEvent
2020-09-08 12:38:58
MHRA/products
https://api.github.com/repos/MHRA/products
closed
DOCUMENTATION - PARs Process Map
EPIC - PARs process
### User want As a internal user I would like to have access to a PARs process map, so that I can understand / amend the process in the future. ### Acceptance Criteria **Customer acceptance criteria** The process map follows the BPMN methodology Agency “House” Best Practice / Style guidelines have been used The process map plots the end to end PARs upload / amend / delete process The process map can be imported into Visual Paradigm (XMI or XML formats - XMI standards preferred) **Technical acceptance criteria** **Data acceptance criteria** **Testing acceptance criteria** **Data - Potential impact** **Size** **Value** **Effort** ### Exit Criteria met - [ ] Backlog - [ ] Discovery - [ ] DUXD - [ ] Development - [ ] Quality Assurance - [ ] Release and Validate
1.0
DOCUMENTATION - PARs Process Map - ### User want As a internal user I would like to have access to a PARs process map, so that I can understand / amend the process in the future. ### Acceptance Criteria **Customer acceptance criteria** The process map follows the BPMN methodology Agency “House” Best Practice / Style guidelines have been used The process map plots the end to end PARs upload / amend / delete process The process map can be imported into Visual Paradigm (XMI or XML formats - XMI standards preferred) **Technical acceptance criteria** **Data acceptance criteria** **Testing acceptance criteria** **Data - Potential impact** **Size** **Value** **Effort** ### Exit Criteria met - [ ] Backlog - [ ] Discovery - [ ] DUXD - [ ] Development - [ ] Quality Assurance - [ ] Release and Validate
process
documentation pars process map user want as a internal user i would like to have access to a pars process map so that i can understand amend the process in the future acceptance criteria customer acceptance criteria the process map follows the bpmn methodology agency “house” best practice style guidelines have been used the process map plots the end to end pars upload amend delete process the process map can be imported into visual paradigm xmi or xml formats xmi standards preferred technical acceptance criteria data acceptance criteria testing acceptance criteria data potential impact size value effort exit criteria met backlog discovery duxd development quality assurance release and validate
1
304,990
23,093,646,524
IssuesEvent
2022-07-26 17:19:38
ericspring08/Awesome-Programming-Languages
https://api.github.com/repos/ericspring08/Awesome-Programming-Languages
closed
Add New Language: V
documentation
# Description Simple language for building maintainable programs. You can learn the entire language by going through the [documentation](https://github.com/vlang/v/blob/master/doc/docs.md) over a weekend, and in most cases there's only one way to do something. This results in simple, readable, and maintainable code. Despite being simple, V gives a lot of power to the developer and can be used in pretty much every field, including systems programming, webdev, gamedev, GUI, mobile, science, embedded, tooling, etc. V is very similar to Go. If you know Go, you already know ≈80% of V.
1.0
Add New Language: V - # Description Simple language for building maintainable programs. You can learn the entire language by going through the [documentation](https://github.com/vlang/v/blob/master/doc/docs.md) over a weekend, and in most cases there's only one way to do something. This results in simple, readable, and maintainable code. Despite being simple, V gives a lot of power to the developer and can be used in pretty much every field, including systems programming, webdev, gamedev, GUI, mobile, science, embedded, tooling, etc. V is very similar to Go. If you know Go, you already know ≈80% of V.
non_process
add new language v description simple language for building maintainable programs you can learn the entire language by going through the over a weekend and in most cases there s only one way to do something this results in simple readable and maintainable code despite being simple v gives a lot of power to the developer and can be used in pretty much every field including systems programming webdev gamedev gui mobile science embedded tooling etc v is very similar to go if you know go you already know ≈ of v
0
10,812
13,609,289,435
IssuesEvent
2020-09-23 04:50:24
googleapis/java-dialogflow
https://api.github.com/repos/googleapis/java-dialogflow
closed
Dependency Dashboard
api: dialogflow type: process
This issue contains a list of Renovate updates and their statuses. ## Open These updates have all been created already. Click a checkbox below to force a retry/rebase of any. - [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-dialogflow-2.x -->chore(deps): update dependency com.google.cloud:google-cloud-dialogflow to v2.2.0 - [ ] <!-- rebase-branch=renovate/com.google.cloud-libraries-bom-10.x -->chore(deps): update dependency com.google.cloud:libraries-bom to v10 --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
1.0
Dependency Dashboard - This issue contains a list of Renovate updates and their statuses. ## Open These updates have all been created already. Click a checkbox below to force a retry/rebase of any. - [ ] <!-- rebase-branch=renovate/com.google.cloud-google-cloud-dialogflow-2.x -->chore(deps): update dependency com.google.cloud:google-cloud-dialogflow to v2.2.0 - [ ] <!-- rebase-branch=renovate/com.google.cloud-libraries-bom-10.x -->chore(deps): update dependency com.google.cloud:libraries-bom to v10 --- - [ ] <!-- manual job -->Check this box to trigger a request for Renovate to run again on this repository
process
dependency dashboard this issue contains a list of renovate updates and their statuses open these updates have all been created already click a checkbox below to force a retry rebase of any chore deps update dependency com google cloud google cloud dialogflow to chore deps update dependency com google cloud libraries bom to check this box to trigger a request for renovate to run again on this repository
1
9,279
12,303,759,590
IssuesEvent
2020-05-11 19:16:15
nextgenhealthcare/connect
https://api.github.com/repos/nextgenhealthcare/connect
closed
Reprocessing a single message by pressing enter causes message to reprocess twice
message reprocess reprocessing twice
In the Channel Messages window, when you highlight a single message and then reprocess it, pressing 'Enter' to select the OK button, two new messages are sent through the channel. Imported Issue. Original Details: Jira Issue Key: MIRTH-2038 Reporter: narupley Created: 2011-12-21T13:01:53.000-0800
2.0
Reprocessing a single message by pressing enter causes message to reprocess twice - In the Channel Messages window, when you highlight a single message and then reprocess it, pressing 'Enter' to select the OK button, two new messages are sent through the channel. Imported Issue. Original Details: Jira Issue Key: MIRTH-2038 Reporter: narupley Created: 2011-12-21T13:01:53.000-0800
process
reprocessing a single message by pressing enter causes message to reprocess twice in the channel messages window when you highlight a single message and then reprocess it pressing enter to select the ok button two new messages are sent through the channel imported issue original details jira issue key mirth reporter narupley created
1
392,142
11,584,044,637
IssuesEvent
2020-02-22 15:03:03
IntegratedTransportPlanning/gcvt
https://api.github.com/repos/IntegratedTransportPlanning/gcvt
opened
Charts are broken on the web server
bug high priority
E.g. http://gcvt.cmcaine.co.uk/api/charts?scenarios=TentRail,DoNothing&variable=Ton&rows=all&width=800&height=500 ``` ######################### ERROR STACKTRACE ######################### IOError: could not spawn setenv(`/root/.julia/artifacts/a42f5c72a500a8683c08a5e818d27a104fc9cd5a/bin/node /root/.julia/artifacts/8e150e7c219ecfd8b18fb27ff830dc6b3628bf86/vl2vg.js`; dir="/root/.julia/artifacts/8e150e7c219ecfd8b18fb27ff830dc6b3628bf86"): no such file or directory (ENOENT) ``` Node does genuinely not exist at that location (or anywhere else) so presumably a package somewhere hasn't installed it properly. Handy debugging step on the web server: `docker exec -it $(docker ps | grep src_julia | cut -d' ' -f1) "/bin/bash"` dumps you in a terminal in the Julia container.
1.0
Charts are broken on the web server - E.g. http://gcvt.cmcaine.co.uk/api/charts?scenarios=TentRail,DoNothing&variable=Ton&rows=all&width=800&height=500 ``` ######################### ERROR STACKTRACE ######################### IOError: could not spawn setenv(`/root/.julia/artifacts/a42f5c72a500a8683c08a5e818d27a104fc9cd5a/bin/node /root/.julia/artifacts/8e150e7c219ecfd8b18fb27ff830dc6b3628bf86/vl2vg.js`; dir="/root/.julia/artifacts/8e150e7c219ecfd8b18fb27ff830dc6b3628bf86"): no such file or directory (ENOENT) ``` Node does genuinely not exist at that location (or anywhere else) so presumably a package somewhere hasn't installed it properly. Handy debugging step on the web server: `docker exec -it $(docker ps | grep src_julia | cut -d' ' -f1) "/bin/bash"` dumps you in a terminal in the Julia container.
non_process
charts are broken on the web server e g error stacktrace ioerror could not spawn setenv root julia artifacts bin node root julia artifacts js dir root julia artifacts no such file or directory enoent node does genuinely not exist at that location or anywhere else so presumably a package somewhere hasn t installed it properly handy debugging step on the web server docker exec it docker ps grep src julia cut d bin bash dumps you in a terminal in the julia container
0
95,101
11,954,043,811
IssuesEvent
2020-04-03 22:21:58
trisagion-games/Convergence-game
https://api.github.com/repos/trisagion-games/Convergence-game
closed
Tutorial Design (Start of Game with Console & Room 1)
level-design
Introducing idea of using a Terminal and Hacking, as well as Rooms and Keys
1.0
Tutorial Design (Start of Game with Console & Room 1) - Introducing idea of using a Terminal and Hacking, as well as Rooms and Keys
non_process
tutorial design start of game with console room introducing idea of using a terminal and hacking as well as rooms and keys
0
5,857
8,680,781,519
IssuesEvent
2018-12-01 14:07:08
bitshares/bitshares-community-ui
https://api.github.com/repos/bitshares/bitshares-community-ui
closed
Improvement on Login via secure key screen
Login P1 process
on the Login screen for the 'SECURE KEY' tab after user browses and selects the key file, and 'password' field appears, **hide** 'backup phrase' field.
1.0
Improvement on Login via secure key screen - on the Login screen for the 'SECURE KEY' tab after user browses and selects the key file, and 'password' field appears, **hide** 'backup phrase' field.
process
improvement on login via secure key screen on the login screen for the secure key tab after user browses and selects the key file and password field appears hide backup phrase field
1
37,538
8,308,927,564
IssuesEvent
2018-09-24 02:04:17
ESAPI/esapi-java-legacy
https://api.github.com/repos/ESAPI/esapi-java-legacy
closed
Remove deprecated fields in Encoder interface
Component-Encoder General Code Cleanup Maintainability
Remove all the deprecated fields in the `Encoder` interface which have equivalents in the `EncoderConstants` class. Related to closed GitHub issue #264. Specifically, remove the following deprecated fields: ``` public final static char[] CHAR_LOWERS = EncoderConstants.CHAR_LOWERS; public final static char[] CHAR_UPPERS = EncoderConstants.CHAR_UPPERS; public final static char[] CHAR_DIGITS = EncoderConstants.CHAR_DIGITS; public final static char[] CHAR_SPECIALS = EncoderConstants.CHAR_SPECIALS; public final static char[] CHAR_LETTERS = EncoderConstants.CHAR_LETTERS; public final static char[] CHAR_ALPHANUMERICS = EncoderConstants.CHAR_ALPHANUMERICS; public final static char[] CHAR_PASSWORD_LOWERS = EncoderConstants.CHAR_PASSWORD_LOWERS; public final static char[] CHAR_PASSWORD_UPPERS = EncoderConstants.CHAR_PASSWORD_UPPERS; public final static char[] CHAR_PASSWORD_DIGITS = EncoderConstants.CHAR_PASSWORD_DIGITS; public final static char[] CHAR_PASSWORD_SPECIALS = EncoderConstants.CHAR_PASSWORD_SPECIALS; public final static char[] CHAR_PASSWORD_LETTERS = EncoderConstants.CHAR_PASSWORD_LETTERS; ```
2.0
Remove deprecated fields in Encoder interface - Remove all the deprecated fields in the `Encoder` interface which have equivalents in the `EncoderConstants` class. Related to closed GitHub issue #264. Specifically, remove the following deprecated fields: ``` public final static char[] CHAR_LOWERS = EncoderConstants.CHAR_LOWERS; public final static char[] CHAR_UPPERS = EncoderConstants.CHAR_UPPERS; public final static char[] CHAR_DIGITS = EncoderConstants.CHAR_DIGITS; public final static char[] CHAR_SPECIALS = EncoderConstants.CHAR_SPECIALS; public final static char[] CHAR_LETTERS = EncoderConstants.CHAR_LETTERS; public final static char[] CHAR_ALPHANUMERICS = EncoderConstants.CHAR_ALPHANUMERICS; public final static char[] CHAR_PASSWORD_LOWERS = EncoderConstants.CHAR_PASSWORD_LOWERS; public final static char[] CHAR_PASSWORD_UPPERS = EncoderConstants.CHAR_PASSWORD_UPPERS; public final static char[] CHAR_PASSWORD_DIGITS = EncoderConstants.CHAR_PASSWORD_DIGITS; public final static char[] CHAR_PASSWORD_SPECIALS = EncoderConstants.CHAR_PASSWORD_SPECIALS; public final static char[] CHAR_PASSWORD_LETTERS = EncoderConstants.CHAR_PASSWORD_LETTERS; ```
non_process
remove deprecated fields in encoder interface remove all the deprecated fields in the encoder interface which have equivalents in the encoderconstants class related to closed github issue specifically remove the following deprecated fields public final static char char lowers encoderconstants char lowers public final static char char uppers encoderconstants char uppers public final static char char digits encoderconstants char digits public final static char char specials encoderconstants char specials public final static char char letters encoderconstants char letters public final static char char alphanumerics encoderconstants char alphanumerics public final static char char password lowers encoderconstants char password lowers public final static char char password uppers encoderconstants char password uppers public final static char char password digits encoderconstants char password digits public final static char char password specials encoderconstants char password specials public final static char char password letters encoderconstants char password letters
0
819,036
30,717,501,421
IssuesEvent
2023-07-27 13:55:20
IRPTeam/IRP
https://api.github.com/repos/IRPTeam/IRP
closed
Negative stock control
Priority
Логика проверки остатков ActualStock по каждому документу меняется с цикла с запросом остатков по каждому документу на один запрос остатков и оборотов в разрезе регистратора При проверке остатков по FreeStock проверяем только конечный остаток (без ограничения по дате). Проверку по каждому документу убираем.
1.0
Negative stock control - Логика проверки остатков ActualStock по каждому документу меняется с цикла с запросом остатков по каждому документу на один запрос остатков и оборотов в разрезе регистратора При проверке остатков по FreeStock проверяем только конечный остаток (без ограничения по дате). Проверку по каждому документу убираем.
non_process
negative stock control логика проверки остатков actualstock по каждому документу меняется с цикла с запросом остатков по каждому документу на один запрос остатков и оборотов в разрезе регистратора при проверке остатков по freestock проверяем только конечный остаток без ограничения по дате проверку по каждому документу убираем
0
813,469
30,459,073,164
IssuesEvent
2023-07-17 04:38:39
Haidoe/arc
https://api.github.com/repos/Haidoe/arc
opened
Produciton Report Forms missing in Mobile Screen
bug priority-medium style
## Bug Report **Reporter: ❗️** @ksdhir **Describe the bug: ❗️** When production report page is accessed on mobile the sidebar takes the whole page and the forms are missing. **Steps to reproduce: ❗️** 1. Go to '[Home Page'](https://arc-app.ca/home) 2. Click on 'Any existing production' 3. Change the browser window to 'mobile view' 4. See the forms are no longer visible. **Screenshots or Video** ***With Sidebar opened*** ![Screenshot 2023-07-16 at 9 15 38 PM](https://github.com/Haidoe/arc/assets/44022086/31b8fc68-2b5b-473a-b6b9-e54208a2ecad) ***With Sidebar closed*** ![Screenshot 2023-07-16 at 9 16 13 PM](https://github.com/Haidoe/arc/assets/44022086/0783acc6-cd59-4170-9641-b22c2d877561) **Expected behavior: ❗️** The layout should look like this: ![Screenshot 2023-07-16 at 9 33 40 PM](https://github.com/Haidoe/arc/assets/44022086/09c72a8f-f9ae-4dd3-925b-8120bdcb9da5) **Actual behavior: ❗️** The sidebar takes the whole window and somehow the forms are still missing even if the sidebar is closed. **Possible Solution:** If you have any ideas or suggestions on how to fix the bug, please mention them here. **Environment:** - Device & Operating System: - Browser and Version(if applicable): - Jira Ticket(if applicable): - Any other relevant information about your environment. **Additional context:** Add any other context about the problem here. **Follow up checklist: ❗️** - [ ] Add Assignee - [ ] Label priority ( priority-low, priority-medium, priority-high ) - [ ] Label Milestone ( Alpha, Beta ) - [ ] Label Issue Type ( Style, Code, API )
1.0
Produciton Report Forms missing in Mobile Screen - ## Bug Report **Reporter: ❗️** @ksdhir **Describe the bug: ❗️** When production report page is accessed on mobile the sidebar takes the whole page and the forms are missing. **Steps to reproduce: ❗️** 1. Go to '[Home Page'](https://arc-app.ca/home) 2. Click on 'Any existing production' 3. Change the browser window to 'mobile view' 4. See the forms are no longer visible. **Screenshots or Video** ***With Sidebar opened*** ![Screenshot 2023-07-16 at 9 15 38 PM](https://github.com/Haidoe/arc/assets/44022086/31b8fc68-2b5b-473a-b6b9-e54208a2ecad) ***With Sidebar closed*** ![Screenshot 2023-07-16 at 9 16 13 PM](https://github.com/Haidoe/arc/assets/44022086/0783acc6-cd59-4170-9641-b22c2d877561) **Expected behavior: ❗️** The layout should look like this: ![Screenshot 2023-07-16 at 9 33 40 PM](https://github.com/Haidoe/arc/assets/44022086/09c72a8f-f9ae-4dd3-925b-8120bdcb9da5) **Actual behavior: ❗️** The sidebar takes the whole window and somehow the forms are still missing even if the sidebar is closed. **Possible Solution:** If you have any ideas or suggestions on how to fix the bug, please mention them here. **Environment:** - Device & Operating System: - Browser and Version(if applicable): - Jira Ticket(if applicable): - Any other relevant information about your environment. **Additional context:** Add any other context about the problem here. **Follow up checklist: ❗️** - [ ] Add Assignee - [ ] Label priority ( priority-low, priority-medium, priority-high ) - [ ] Label Milestone ( Alpha, Beta ) - [ ] Label Issue Type ( Style, Code, API )
non_process
produciton report forms missing in mobile screen bug report reporter ❗️ ksdhir describe the bug ❗️ when production report page is accessed on mobile the sidebar takes the whole page and the forms are missing steps to reproduce ❗️ go to click on any existing production change the browser window to mobile view see the forms are no longer visible screenshots or video with sidebar opened with sidebar closed expected behavior ❗️ the layout should look like this actual behavior ❗️ the sidebar takes the whole window and somehow the forms are still missing even if the sidebar is closed possible solution if you have any ideas or suggestions on how to fix the bug please mention them here environment device operating system browser and version if applicable jira ticket if applicable any other relevant information about your environment additional context add any other context about the problem here follow up checklist ❗️ add assignee label priority priority low priority medium priority high label milestone alpha beta label issue type style code api
0